var/home/core/zuul-output/0000755000175000017500000000000015117317271014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015117323356015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004162257215117323350017706 0ustar rootrootDec 13 17:16:24 crc systemd[1]: Starting Kubernetes Kubelet... Dec 13 17:16:24 crc restorecon[4695]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:24 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:25 crc restorecon[4695]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 13 17:16:25 crc restorecon[4695]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 13 17:16:25 crc kubenswrapper[4989]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 17:16:25 crc kubenswrapper[4989]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 13 17:16:25 crc kubenswrapper[4989]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 17:16:25 crc kubenswrapper[4989]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 17:16:25 crc kubenswrapper[4989]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 13 17:16:25 crc kubenswrapper[4989]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.826392 4989 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833154 4989 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833186 4989 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833197 4989 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833206 4989 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833215 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833224 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833233 4989 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833249 4989 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833259 4989 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833268 4989 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833276 4989 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833284 4989 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833293 4989 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833301 4989 feature_gate.go:330] unrecognized feature gate: Example Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833311 4989 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833320 4989 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833328 4989 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833337 4989 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833345 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833354 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833362 4989 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833370 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833379 4989 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833387 4989 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833396 4989 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833405 4989 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833413 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833422 4989 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833430 4989 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833441 4989 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833453 4989 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833463 4989 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833472 4989 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833482 4989 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833491 4989 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833500 4989 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833511 4989 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833523 4989 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833535 4989 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833544 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833553 4989 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833562 4989 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833572 4989 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833581 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833589 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833599 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833608 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833618 4989 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833627 4989 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833636 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833647 4989 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833658 4989 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833667 4989 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833676 4989 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833685 4989 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833694 4989 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833703 4989 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833712 4989 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833722 4989 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833731 4989 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833740 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833749 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833759 4989 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833770 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833782 4989 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833824 4989 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833834 4989 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833844 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833852 4989 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833861 4989 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.833870 4989 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834262 4989 flags.go:64] FLAG: --address="0.0.0.0" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834285 4989 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834303 4989 flags.go:64] FLAG: --anonymous-auth="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834316 4989 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834328 4989 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834339 4989 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834351 4989 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834363 4989 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834373 4989 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834383 4989 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834394 4989 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834404 4989 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834414 4989 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834424 4989 flags.go:64] FLAG: --cgroup-root="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834434 4989 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834444 4989 flags.go:64] FLAG: --client-ca-file="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834454 4989 flags.go:64] FLAG: --cloud-config="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834463 4989 flags.go:64] FLAG: --cloud-provider="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834472 4989 flags.go:64] FLAG: --cluster-dns="[]" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834484 4989 flags.go:64] FLAG: --cluster-domain="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834493 4989 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834503 4989 flags.go:64] FLAG: --config-dir="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834512 4989 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834523 4989 flags.go:64] FLAG: --container-log-max-files="5" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834535 4989 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834545 4989 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834554 4989 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834564 4989 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834574 4989 flags.go:64] FLAG: --contention-profiling="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834586 4989 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834634 4989 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834644 4989 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834654 4989 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834665 4989 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834675 4989 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834685 4989 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834695 4989 flags.go:64] FLAG: --enable-load-reader="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834704 4989 flags.go:64] FLAG: --enable-server="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834715 4989 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834727 4989 flags.go:64] FLAG: --event-burst="100" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834737 4989 flags.go:64] FLAG: --event-qps="50" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834747 4989 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834757 4989 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834767 4989 flags.go:64] FLAG: --eviction-hard="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834787 4989 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834822 4989 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834831 4989 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834841 4989 flags.go:64] FLAG: --eviction-soft="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834851 4989 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834861 4989 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834871 4989 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834880 4989 flags.go:64] FLAG: --experimental-mounter-path="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834890 4989 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834900 4989 flags.go:64] FLAG: --fail-swap-on="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834909 4989 flags.go:64] FLAG: --feature-gates="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834921 4989 flags.go:64] FLAG: --file-check-frequency="20s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834931 4989 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834941 4989 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834951 4989 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834961 4989 flags.go:64] FLAG: --healthz-port="10248" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834971 4989 flags.go:64] FLAG: --help="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834981 4989 flags.go:64] FLAG: --hostname-override="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.834991 4989 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835000 4989 flags.go:64] FLAG: --http-check-frequency="20s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835012 4989 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835022 4989 flags.go:64] FLAG: --image-credential-provider-config="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835033 4989 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835043 4989 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835052 4989 flags.go:64] FLAG: --image-service-endpoint="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835062 4989 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835071 4989 flags.go:64] FLAG: --kube-api-burst="100" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835081 4989 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835091 4989 flags.go:64] FLAG: --kube-api-qps="50" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835101 4989 flags.go:64] FLAG: --kube-reserved="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835112 4989 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835121 4989 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835132 4989 flags.go:64] FLAG: --kubelet-cgroups="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835141 4989 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835150 4989 flags.go:64] FLAG: --lock-file="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835160 4989 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835171 4989 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835181 4989 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835195 4989 flags.go:64] FLAG: --log-json-split-stream="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835205 4989 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835214 4989 flags.go:64] FLAG: --log-text-split-stream="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835224 4989 flags.go:64] FLAG: --logging-format="text" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835234 4989 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835244 4989 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835254 4989 flags.go:64] FLAG: --manifest-url="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835263 4989 flags.go:64] FLAG: --manifest-url-header="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835275 4989 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835285 4989 flags.go:64] FLAG: --max-open-files="1000000" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835297 4989 flags.go:64] FLAG: --max-pods="110" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835308 4989 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835318 4989 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835328 4989 flags.go:64] FLAG: --memory-manager-policy="None" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835337 4989 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835348 4989 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835357 4989 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835367 4989 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835387 4989 flags.go:64] FLAG: --node-status-max-images="50" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835397 4989 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835407 4989 flags.go:64] FLAG: --oom-score-adj="-999" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835417 4989 flags.go:64] FLAG: --pod-cidr="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835427 4989 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835441 4989 flags.go:64] FLAG: --pod-manifest-path="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835450 4989 flags.go:64] FLAG: --pod-max-pids="-1" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835461 4989 flags.go:64] FLAG: --pods-per-core="0" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835470 4989 flags.go:64] FLAG: --port="10250" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835480 4989 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835491 4989 flags.go:64] FLAG: --provider-id="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835500 4989 flags.go:64] FLAG: --qos-reserved="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835510 4989 flags.go:64] FLAG: --read-only-port="10255" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835520 4989 flags.go:64] FLAG: --register-node="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835530 4989 flags.go:64] FLAG: --register-schedulable="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835539 4989 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835556 4989 flags.go:64] FLAG: --registry-burst="10" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835565 4989 flags.go:64] FLAG: --registry-qps="5" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835575 4989 flags.go:64] FLAG: --reserved-cpus="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835584 4989 flags.go:64] FLAG: --reserved-memory="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835643 4989 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835654 4989 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835681 4989 flags.go:64] FLAG: --rotate-certificates="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835692 4989 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835702 4989 flags.go:64] FLAG: --runonce="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835712 4989 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835722 4989 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835731 4989 flags.go:64] FLAG: --seccomp-default="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835741 4989 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835751 4989 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835761 4989 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835771 4989 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835781 4989 flags.go:64] FLAG: --storage-driver-password="root" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835817 4989 flags.go:64] FLAG: --storage-driver-secure="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835827 4989 flags.go:64] FLAG: --storage-driver-table="stats" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835837 4989 flags.go:64] FLAG: --storage-driver-user="root" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835846 4989 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835856 4989 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835866 4989 flags.go:64] FLAG: --system-cgroups="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835875 4989 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835891 4989 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835900 4989 flags.go:64] FLAG: --tls-cert-file="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835909 4989 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835922 4989 flags.go:64] FLAG: --tls-min-version="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835932 4989 flags.go:64] FLAG: --tls-private-key-file="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835942 4989 flags.go:64] FLAG: --topology-manager-policy="none" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835953 4989 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835964 4989 flags.go:64] FLAG: --topology-manager-scope="container" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835973 4989 flags.go:64] FLAG: --v="2" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835986 4989 flags.go:64] FLAG: --version="false" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.835998 4989 flags.go:64] FLAG: --vmodule="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.836010 4989 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.836021 4989 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836237 4989 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836249 4989 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836260 4989 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836269 4989 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836279 4989 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836288 4989 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836297 4989 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836307 4989 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836316 4989 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836325 4989 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836333 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836345 4989 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836356 4989 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836366 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836375 4989 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836384 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836392 4989 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836401 4989 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836409 4989 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836420 4989 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836429 4989 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836437 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836446 4989 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836455 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836464 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836472 4989 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836480 4989 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836489 4989 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836498 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836509 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836521 4989 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836532 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836541 4989 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836552 4989 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836561 4989 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836571 4989 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836579 4989 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836588 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836598 4989 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836607 4989 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836615 4989 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836623 4989 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836631 4989 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836640 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836648 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836657 4989 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836665 4989 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836673 4989 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836682 4989 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836690 4989 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836699 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836707 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836715 4989 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836725 4989 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836733 4989 feature_gate.go:330] unrecognized feature gate: Example Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836742 4989 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836750 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836758 4989 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836767 4989 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836776 4989 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836784 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836819 4989 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836828 4989 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836837 4989 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836846 4989 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836855 4989 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836865 4989 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836874 4989 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836883 4989 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836894 4989 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.836905 4989 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.836918 4989 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.847412 4989 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.847460 4989 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847589 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847610 4989 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847620 4989 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847631 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847641 4989 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847650 4989 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847658 4989 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847667 4989 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847676 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847685 4989 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847693 4989 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847701 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847709 4989 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847718 4989 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847726 4989 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847735 4989 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847742 4989 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847750 4989 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847758 4989 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847766 4989 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847774 4989 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847782 4989 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847818 4989 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847831 4989 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847840 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847851 4989 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847861 4989 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847869 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847877 4989 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847885 4989 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847894 4989 feature_gate.go:330] unrecognized feature gate: Example Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847904 4989 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847913 4989 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847921 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847930 4989 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847937 4989 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847945 4989 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847953 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847964 4989 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847974 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847983 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.847993 4989 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848002 4989 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848012 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848023 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848032 4989 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848042 4989 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848052 4989 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848062 4989 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848071 4989 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848080 4989 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848089 4989 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848098 4989 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848106 4989 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848116 4989 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848125 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848134 4989 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848143 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848154 4989 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848165 4989 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848175 4989 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848183 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848193 4989 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848208 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848228 4989 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848240 4989 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848252 4989 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848261 4989 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848272 4989 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848282 4989 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848292 4989 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.848308 4989 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848571 4989 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848588 4989 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848596 4989 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848606 4989 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848616 4989 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848624 4989 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848634 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848643 4989 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848651 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848660 4989 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848668 4989 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848678 4989 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848690 4989 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848698 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848708 4989 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848716 4989 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848726 4989 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848735 4989 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848743 4989 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848751 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848760 4989 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848768 4989 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848776 4989 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848784 4989 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848829 4989 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848839 4989 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848847 4989 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848855 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848863 4989 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848871 4989 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848879 4989 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848890 4989 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848900 4989 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848909 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848918 4989 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848926 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848934 4989 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848942 4989 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848950 4989 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848958 4989 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848966 4989 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848976 4989 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848986 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.848995 4989 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849003 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849014 4989 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849023 4989 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849033 4989 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849041 4989 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849049 4989 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849058 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849065 4989 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849073 4989 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849081 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849089 4989 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849097 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849105 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849113 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849120 4989 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849129 4989 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849137 4989 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849145 4989 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849152 4989 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849160 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849168 4989 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849175 4989 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849184 4989 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849191 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849200 4989 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849209 4989 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.849219 4989 feature_gate.go:330] unrecognized feature gate: Example Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.849242 4989 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.849815 4989 server.go:940] "Client rotation is on, will bootstrap in background" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.854653 4989 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.854896 4989 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.855913 4989 server.go:997] "Starting client certificate rotation" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.855957 4989 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.856497 4989 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-27 11:43:07.647828218 +0000 UTC Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.856643 4989 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.863470 4989 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 13 17:16:25 crc kubenswrapper[4989]: E1213 17:16:25.865355 4989 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.868495 4989 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.885012 4989 log.go:25] "Validated CRI v1 runtime API" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.916712 4989 log.go:25] "Validated CRI v1 image API" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.917883 4989 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.919971 4989 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-13-17-12-07-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.920020 4989 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.946688 4989 manager.go:217] Machine: {Timestamp:2025-12-13 17:16:25.943396914 +0000 UTC m=+0.549844092 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:364e0ead-9bda-403b-8614-341ecd5845fc BootID:9e98691d-d678-4159-b20c-43488bb99dd0 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:43:60:39 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:43:60:39 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:c8:0f:9e Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:f3:6f:07 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:e2:21:86 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:9c:f4:ee Speed:-1 Mtu:1496} {Name:eth10 MacAddress:36:cb:b9:f4:bc:c5 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:7e:73:cd:d7:66:a4 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.946989 4989 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.947186 4989 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.947464 4989 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.947671 4989 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.947710 4989 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.948031 4989 topology_manager.go:138] "Creating topology manager with none policy" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.948044 4989 container_manager_linux.go:303] "Creating device plugin manager" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.948185 4989 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.948220 4989 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.948552 4989 state_mem.go:36] "Initialized new in-memory state store" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.948832 4989 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.949542 4989 kubelet.go:418] "Attempting to sync node with API server" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.949564 4989 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.949616 4989 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.949631 4989 kubelet.go:324] "Adding apiserver pod source" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.949647 4989 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.951650 4989 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.952055 4989 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954140 4989 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.954353 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Dec 13 17:16:25 crc kubenswrapper[4989]: E1213 17:16:25.954418 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.954397 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Dec 13 17:16:25 crc kubenswrapper[4989]: E1213 17:16:25.954469 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954707 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954734 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954744 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954754 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954769 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954778 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954805 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954820 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954831 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954842 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954855 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.954864 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.959196 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.959708 4989 server.go:1280] "Started kubelet" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.960664 4989 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.960666 4989 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.961121 4989 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.961343 4989 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Dec 13 17:16:25 crc systemd[1]: Started Kubernetes Kubelet. Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.962160 4989 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.962198 4989 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.962246 4989 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 01:39:56.039299832 +0000 UTC Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.962303 4989 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 32h23m30.07700204s for next certificate rotation Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.962442 4989 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.962453 4989 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.962532 4989 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.962825 4989 server.go:460] "Adding debug handlers to kubelet server" Dec 13 17:16:25 crc kubenswrapper[4989]: E1213 17:16:25.962358 4989 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 13 17:16:25 crc kubenswrapper[4989]: E1213 17:16:25.963805 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="200ms" Dec 13 17:16:25 crc kubenswrapper[4989]: W1213 17:16:25.967067 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Dec 13 17:16:25 crc kubenswrapper[4989]: E1213 17:16:25.967152 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.968345 4989 factory.go:55] Registering systemd factory Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.968375 4989 factory.go:221] Registration of the systemd container factory successfully Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.969329 4989 factory.go:153] Registering CRI-O factory Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.969359 4989 factory.go:221] Registration of the crio container factory successfully Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.969424 4989 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.969451 4989 factory.go:103] Registering Raw factory Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.969466 4989 manager.go:1196] Started watching for new ooms in manager Dec 13 17:16:25 crc kubenswrapper[4989]: E1213 17:16:25.966329 4989 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.47:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1880d5d9f1265a39 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-13 17:16:25.959676473 +0000 UTC m=+0.566123631,LastTimestamp:2025-12-13 17:16:25.959676473 +0000 UTC m=+0.566123631,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.975474 4989 manager.go:319] Starting recovery of all containers Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.979297 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.979532 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.979644 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.979752 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.979892 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.980021 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.980131 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.980242 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.980353 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.980463 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.980578 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.980700 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.980844 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.980964 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.981072 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.981186 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.981309 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.981420 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.981529 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.981639 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.981741 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.981901 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.982019 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.982128 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.982233 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.982337 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.982457 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.982566 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.982687 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.982818 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.982934 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.983116 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.983236 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.983460 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.983573 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.983676 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.983812 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.983946 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.984059 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.984173 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.984276 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.984378 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.984494 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.984606 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.984728 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.984931 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.985040 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.985169 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.985278 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.985387 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.985495 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.985605 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.985725 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.985891 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.986020 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.986166 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.987292 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.987465 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.987582 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.987685 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.987813 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.987924 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.988024 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.988143 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.988242 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.988336 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.988428 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.988519 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.988611 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.988708 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.988830 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.988930 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.989025 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.989117 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.989226 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.989320 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.989420 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.989512 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.989607 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.989711 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.989829 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.989930 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.990026 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.990120 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.990215 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.990322 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.990422 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.990535 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.990632 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.990723 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.990850 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.990953 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.991046 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.991137 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.991229 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.991364 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.991465 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.991560 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.991652 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.991751 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.991916 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.992025 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.992128 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.992240 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.992344 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.992450 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.992568 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.992673 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.992776 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.992913 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.993011 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.993118 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.993204 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.993293 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.993375 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.993847 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.993981 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.994109 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.991720 4989 manager.go:324] Recovery completed Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995209 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995367 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995384 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995396 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995407 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995418 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995428 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995441 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995451 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995462 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995471 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995482 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995492 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995501 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995512 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995522 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995535 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995546 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995556 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995566 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995575 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995586 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995596 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995605 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995615 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995625 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995635 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995647 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995656 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995665 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995674 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995684 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995693 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995703 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995712 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995721 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995730 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995740 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995748 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995759 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995768 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995777 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995787 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995819 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995830 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995841 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995851 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995862 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995873 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995883 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995893 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995904 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995913 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995923 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995934 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995944 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995955 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995966 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.995976 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996530 4989 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996556 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996568 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996578 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996588 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996598 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996607 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996616 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996625 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996636 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996644 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996654 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996663 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996673 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996683 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996692 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996702 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996711 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996721 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996729 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996739 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996748 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996761 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996770 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996779 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996807 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996820 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996829 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996838 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996847 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996857 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996865 4989 reconstruct.go:97] "Volume reconstruction finished" Dec 13 17:16:25 crc kubenswrapper[4989]: I1213 17:16:25.996872 4989 reconciler.go:26] "Reconciler: start to sync state" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.007058 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.008712 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.008755 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.008767 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.010986 4989 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.011917 4989 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.011942 4989 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.011963 4989 state_mem.go:36] "Initialized new in-memory state store" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.013253 4989 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.013286 4989 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.013309 4989 kubelet.go:2335] "Starting kubelet main sync loop" Dec 13 17:16:26 crc kubenswrapper[4989]: E1213 17:16:26.013350 4989 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 13 17:16:26 crc kubenswrapper[4989]: W1213 17:16:26.015728 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Dec 13 17:16:26 crc kubenswrapper[4989]: E1213 17:16:26.015778 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.027397 4989 policy_none.go:49] "None policy: Start" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.028071 4989 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.028101 4989 state_mem.go:35] "Initializing new in-memory state store" Dec 13 17:16:26 crc kubenswrapper[4989]: E1213 17:16:26.063159 4989 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.077981 4989 manager.go:334] "Starting Device Plugin manager" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.078035 4989 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.078049 4989 server.go:79] "Starting device plugin registration server" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.078476 4989 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.078491 4989 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.078638 4989 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.078771 4989 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.078780 4989 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 13 17:16:26 crc kubenswrapper[4989]: E1213 17:16:26.087851 4989 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.113502 4989 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.113656 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.114758 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.114817 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.114830 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.114995 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.115123 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.115153 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.115688 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.115699 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.115722 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.115722 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.115732 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.115739 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.115854 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.116083 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.116134 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.116407 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.116429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.116441 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.116528 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.116654 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.116683 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.117210 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.117230 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.117238 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.117281 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.117295 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.117304 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.118031 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.118058 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.118070 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.118196 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.118349 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.118376 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.119287 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.119307 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.119338 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.119349 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.119312 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.119510 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.119642 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.119671 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.120331 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.120355 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.120365 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: E1213 17:16:26.164870 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="400ms" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.179293 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.180281 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.180315 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.180324 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.180343 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 17:16:26 crc kubenswrapper[4989]: E1213 17:16:26.180630 4989 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198489 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198512 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198528 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198545 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198566 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198601 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198633 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198653 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198668 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198700 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198730 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198761 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198807 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198831 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.198850 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300110 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300153 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300173 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300189 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300206 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300225 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300241 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300259 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300275 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300290 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300306 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300321 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300325 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300351 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300390 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300398 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300370 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300415 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300336 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300364 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300361 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300436 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300446 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300448 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300447 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300496 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300493 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300577 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300560 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.300542 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.381370 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.382763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.382834 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.382845 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.382872 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 17:16:26 crc kubenswrapper[4989]: E1213 17:16:26.383392 4989 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.447077 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.454442 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.472628 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: W1213 17:16:26.489069 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-782b0a98e628553fb0dc8938e2b6400b4d4d95a7d464125da10a96523d791e9e WatchSource:0}: Error finding container 782b0a98e628553fb0dc8938e2b6400b4d4d95a7d464125da10a96523d791e9e: Status 404 returned error can't find the container with id 782b0a98e628553fb0dc8938e2b6400b4d4d95a7d464125da10a96523d791e9e Dec 13 17:16:26 crc kubenswrapper[4989]: W1213 17:16:26.495810 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-dd5f20b93c766350b4cf5b42d36cb2685f6b378852f5dfad9e9f8218d36f70ca WatchSource:0}: Error finding container dd5f20b93c766350b4cf5b42d36cb2685f6b378852f5dfad9e9f8218d36f70ca: Status 404 returned error can't find the container with id dd5f20b93c766350b4cf5b42d36cb2685f6b378852f5dfad9e9f8218d36f70ca Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.495954 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.501541 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 17:16:26 crc kubenswrapper[4989]: W1213 17:16:26.524117 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-b08e62bf92bfc648eb9a4ddd390e91cd086bfc9bd4d538415bd6e23fafdcf4a2 WatchSource:0}: Error finding container b08e62bf92bfc648eb9a4ddd390e91cd086bfc9bd4d538415bd6e23fafdcf4a2: Status 404 returned error can't find the container with id b08e62bf92bfc648eb9a4ddd390e91cd086bfc9bd4d538415bd6e23fafdcf4a2 Dec 13 17:16:26 crc kubenswrapper[4989]: E1213 17:16:26.566715 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="800ms" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.784221 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.785462 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.785503 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.785512 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.785539 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 17:16:26 crc kubenswrapper[4989]: E1213 17:16:26.786000 4989 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Dec 13 17:16:26 crc kubenswrapper[4989]: E1213 17:16:26.900160 4989 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.47:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1880d5d9f1265a39 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-13 17:16:25.959676473 +0000 UTC m=+0.566123631,LastTimestamp:2025-12-13 17:16:25.959676473 +0000 UTC m=+0.566123631,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 13 17:16:26 crc kubenswrapper[4989]: W1213 17:16:26.906765 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Dec 13 17:16:26 crc kubenswrapper[4989]: E1213 17:16:26.907415 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Dec 13 17:16:26 crc kubenswrapper[4989]: I1213 17:16:26.963024 4989 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.018611 4989 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="386f0a89a932893d541d90c3daec49677e1ae574c5bc75138187d7f67b0b0c9a" exitCode=0 Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.018700 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"386f0a89a932893d541d90c3daec49677e1ae574c5bc75138187d7f67b0b0c9a"} Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.018837 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"bcc6f9d136471e819bca5fd173303423662c83c81ce05227a4fce6ba7894499a"} Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.018920 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.020124 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.020167 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.020180 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.021806 4989 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d" exitCode=0 Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.021836 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d"} Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.021882 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a6b48c696f7d516f2aaed24137de535b782ae7372926433aa55343eac4381ffc"} Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.021947 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.022583 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.022609 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.022618 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.024454 4989 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26" exitCode=0 Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.024564 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26"} Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.024601 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b08e62bf92bfc648eb9a4ddd390e91cd086bfc9bd4d538415bd6e23fafdcf4a2"} Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.024780 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.025851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.025889 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.025901 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.026871 4989 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008" exitCode=0 Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.026940 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008"} Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.026967 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dd5f20b93c766350b4cf5b42d36cb2685f6b378852f5dfad9e9f8218d36f70ca"} Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.027099 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.027953 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.028678 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.028710 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.028680 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.028721 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.028736 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.028748 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.029011 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d"} Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.029040 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"782b0a98e628553fb0dc8938e2b6400b4d4d95a7d464125da10a96523d791e9e"} Dec 13 17:16:27 crc kubenswrapper[4989]: W1213 17:16:27.088841 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Dec 13 17:16:27 crc kubenswrapper[4989]: E1213 17:16:27.088936 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Dec 13 17:16:27 crc kubenswrapper[4989]: W1213 17:16:27.106389 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Dec 13 17:16:27 crc kubenswrapper[4989]: E1213 17:16:27.106488 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Dec 13 17:16:27 crc kubenswrapper[4989]: W1213 17:16:27.127914 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Dec 13 17:16:27 crc kubenswrapper[4989]: E1213 17:16:27.127986 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Dec 13 17:16:27 crc kubenswrapper[4989]: E1213 17:16:27.367601 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="1.6s" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.586839 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.587974 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.588011 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.588020 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.588043 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 17:16:27 crc kubenswrapper[4989]: E1213 17:16:27.588486 4989 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.892156 4989 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 13 17:16:27 crc kubenswrapper[4989]: E1213 17:16:27.893695 4989 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Dec 13 17:16:27 crc kubenswrapper[4989]: I1213 17:16:27.962876 4989 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.034137 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.034175 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.034185 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.034258 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.034994 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.035017 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.035026 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.036868 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.036888 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.036900 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.036910 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.038886 4989 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014" exitCode=0 Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.038929 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.038988 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.039600 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.039616 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.039626 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.040698 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.040722 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.040733 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.040786 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.041388 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.041409 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.041429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.045953 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"961b7ddbae6e098a3f15adc1253a50fc94fcd98c362f248e976570a9a391769f"} Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.046153 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.047260 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.047284 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:28 crc kubenswrapper[4989]: I1213 17:16:28.047293 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.051119 4989 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9" exitCode=0 Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.051215 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9"} Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.051380 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.052318 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.052370 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.052385 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.058620 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db"} Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.058635 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.058729 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.059469 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.059505 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.059519 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.060047 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.060085 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.060097 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.188620 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.189834 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.189882 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.189897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.189923 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 17:16:29 crc kubenswrapper[4989]: I1213 17:16:29.514963 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.065157 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.065204 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.065143 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516"} Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.065272 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e"} Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.065324 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee"} Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.065335 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb"} Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.065343 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379"} Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.065378 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.065902 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.065942 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.065954 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.066482 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.066509 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.066518 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.685642 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.685850 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.687027 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.687069 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:30 crc kubenswrapper[4989]: I1213 17:16:30.687080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:31 crc kubenswrapper[4989]: I1213 17:16:31.067677 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:31 crc kubenswrapper[4989]: I1213 17:16:31.067698 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 13 17:16:31 crc kubenswrapper[4989]: I1213 17:16:31.067742 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:31 crc kubenswrapper[4989]: I1213 17:16:31.068928 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:31 crc kubenswrapper[4989]: I1213 17:16:31.068971 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:31 crc kubenswrapper[4989]: I1213 17:16:31.068988 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:31 crc kubenswrapper[4989]: I1213 17:16:31.069025 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:31 crc kubenswrapper[4989]: I1213 17:16:31.069077 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:31 crc kubenswrapper[4989]: I1213 17:16:31.069107 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:31 crc kubenswrapper[4989]: I1213 17:16:31.276877 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:32 crc kubenswrapper[4989]: I1213 17:16:32.068341 4989 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 13 17:16:32 crc kubenswrapper[4989]: I1213 17:16:32.070018 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 13 17:16:32 crc kubenswrapper[4989]: I1213 17:16:32.070068 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:32 crc kubenswrapper[4989]: I1213 17:16:32.071089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:32 crc kubenswrapper[4989]: I1213 17:16:32.071146 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:32 crc kubenswrapper[4989]: I1213 17:16:32.071164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:33 crc kubenswrapper[4989]: I1213 17:16:33.369249 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 13 17:16:33 crc kubenswrapper[4989]: I1213 17:16:33.369451 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:33 crc kubenswrapper[4989]: I1213 17:16:33.370619 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:33 crc kubenswrapper[4989]: I1213 17:16:33.370665 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:33 crc kubenswrapper[4989]: I1213 17:16:33.370678 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:33 crc kubenswrapper[4989]: I1213 17:16:33.525076 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:33 crc kubenswrapper[4989]: I1213 17:16:33.525360 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:33 crc kubenswrapper[4989]: I1213 17:16:33.527008 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:33 crc kubenswrapper[4989]: I1213 17:16:33.527070 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:33 crc kubenswrapper[4989]: I1213 17:16:33.527087 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:35 crc kubenswrapper[4989]: I1213 17:16:35.077465 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:35 crc kubenswrapper[4989]: I1213 17:16:35.077725 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:35 crc kubenswrapper[4989]: I1213 17:16:35.079487 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:35 crc kubenswrapper[4989]: I1213 17:16:35.079646 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:35 crc kubenswrapper[4989]: I1213 17:16:35.079674 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:36 crc kubenswrapper[4989]: I1213 17:16:36.014195 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:36 crc kubenswrapper[4989]: I1213 17:16:36.014498 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:36 crc kubenswrapper[4989]: I1213 17:16:36.016390 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:36 crc kubenswrapper[4989]: I1213 17:16:36.016458 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:36 crc kubenswrapper[4989]: I1213 17:16:36.016478 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:36 crc kubenswrapper[4989]: E1213 17:16:36.088008 4989 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 13 17:16:36 crc kubenswrapper[4989]: I1213 17:16:36.528380 4989 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 13 17:16:36 crc kubenswrapper[4989]: I1213 17:16:36.528499 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.049626 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.049870 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.051403 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.051435 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.051447 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.054970 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.083752 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.085113 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.085200 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.085345 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.090855 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:37 crc kubenswrapper[4989]: I1213 17:16:37.257421 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.085505 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.089464 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.089541 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.089573 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.842719 4989 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.842860 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.849130 4989 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.849225 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.878855 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.879138 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.880852 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.880932 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:38 crc kubenswrapper[4989]: I1213 17:16:38.880965 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.089087 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.090534 4989 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db" exitCode=255 Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.090652 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.091098 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db"} Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.091191 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.091688 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.091712 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.091720 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.092104 4989 scope.go:117] "RemoveContainer" containerID="a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db" Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.092461 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.092496 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.092508 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.526488 4989 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]log ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]etcd ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/generic-apiserver-start-informers ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/priority-and-fairness-filter ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/start-apiextensions-informers ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/start-apiextensions-controllers ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/crd-informer-synced ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/start-system-namespaces-controller ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 13 17:16:39 crc kubenswrapper[4989]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 13 17:16:39 crc kubenswrapper[4989]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/bootstrap-controller ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/start-kube-aggregator-informers ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/apiservice-registration-controller ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/apiservice-discovery-controller ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]autoregister-completion ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/apiservice-openapi-controller ok Dec 13 17:16:39 crc kubenswrapper[4989]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 13 17:16:39 crc kubenswrapper[4989]: livez check failed Dec 13 17:16:39 crc kubenswrapper[4989]: I1213 17:16:39.526872 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:16:40 crc kubenswrapper[4989]: I1213 17:16:40.096076 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 13 17:16:40 crc kubenswrapper[4989]: I1213 17:16:40.098730 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986"} Dec 13 17:16:40 crc kubenswrapper[4989]: I1213 17:16:40.098902 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:40 crc kubenswrapper[4989]: I1213 17:16:40.100100 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:40 crc kubenswrapper[4989]: I1213 17:16:40.100166 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:40 crc kubenswrapper[4989]: I1213 17:16:40.100181 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:43 crc kubenswrapper[4989]: E1213 17:16:43.832863 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.834648 4989 trace.go:236] Trace[1703123745]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Dec-2025 17:16:29.716) (total time: 14117ms): Dec 13 17:16:43 crc kubenswrapper[4989]: Trace[1703123745]: ---"Objects listed" error: 14117ms (17:16:43.834) Dec 13 17:16:43 crc kubenswrapper[4989]: Trace[1703123745]: [14.117668337s] [14.117668337s] END Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.834684 4989 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.835127 4989 trace.go:236] Trace[1446215224]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Dec-2025 17:16:29.034) (total time: 14800ms): Dec 13 17:16:43 crc kubenswrapper[4989]: Trace[1446215224]: ---"Objects listed" error: 14800ms (17:16:43.835) Dec 13 17:16:43 crc kubenswrapper[4989]: Trace[1446215224]: [14.800618778s] [14.800618778s] END Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.835160 4989 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.835939 4989 trace.go:236] Trace[130164173]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Dec-2025 17:16:29.499) (total time: 14336ms): Dec 13 17:16:43 crc kubenswrapper[4989]: Trace[130164173]: ---"Objects listed" error: 14336ms (17:16:43.835) Dec 13 17:16:43 crc kubenswrapper[4989]: Trace[130164173]: [14.336783482s] [14.336783482s] END Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.835956 4989 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.836871 4989 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.837098 4989 trace.go:236] Trace[1554575482]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Dec-2025 17:16:28.856) (total time: 14980ms): Dec 13 17:16:43 crc kubenswrapper[4989]: Trace[1554575482]: ---"Objects listed" error: 14980ms (17:16:43.836) Dec 13 17:16:43 crc kubenswrapper[4989]: Trace[1554575482]: [14.980992051s] [14.980992051s] END Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.837135 4989 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 13 17:16:43 crc kubenswrapper[4989]: E1213 17:16:43.838345 4989 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.843406 4989 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.869321 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.875241 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.959940 4989 apiserver.go:52] "Watching apiserver" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.964943 4989 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.965300 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.965609 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.965825 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.966098 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:43 crc kubenswrapper[4989]: E1213 17:16:43.966203 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.966228 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.966261 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 17:16:43 crc kubenswrapper[4989]: E1213 17:16:43.966268 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.966627 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:43 crc kubenswrapper[4989]: E1213 17:16:43.966717 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.967714 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.968005 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.968121 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.968279 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.968587 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.968675 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.968849 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.969926 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 13 17:16:43 crc kubenswrapper[4989]: I1213 17:16:43.972128 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.008277 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.021480 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.030852 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.038639 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.047878 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.057287 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.064042 4989 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.064206 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.074081 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.117237 4989 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.138917 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.138957 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.138977 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139001 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139025 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139131 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139152 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139169 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139219 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139236 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139251 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139292 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139313 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139331 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139351 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139387 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139382 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139410 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139427 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139483 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139511 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139537 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139560 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139587 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139609 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139611 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139631 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139650 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139658 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139699 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139722 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139731 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139839 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139846 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139875 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139968 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140009 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140049 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140085 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140120 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140254 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140290 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140323 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140357 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140387 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140419 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140492 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140524 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140545 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140565 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140587 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140608 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140632 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140654 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140677 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140699 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140750 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140773 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140813 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140836 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140857 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140882 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140915 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140937 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140960 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140982 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141003 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141026 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141050 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141072 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141097 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141119 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141142 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141164 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141216 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141240 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141261 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141286 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141308 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141335 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141360 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141384 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141406 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141428 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.139975 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141451 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141456 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140011 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140081 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140123 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140124 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140214 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140433 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140430 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140595 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140600 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140665 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140748 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140769 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140824 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140832 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140884 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.140891 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141017 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141123 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141122 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141155 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141160 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141350 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141424 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141423 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141466 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141473 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141668 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141679 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141699 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141721 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141745 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141760 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141764 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141778 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141810 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141827 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141844 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141860 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141879 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141898 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141915 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141937 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141956 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141972 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141987 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142013 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142028 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142048 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142066 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142085 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142099 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142115 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142130 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142144 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142161 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142176 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142192 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142206 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142227 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142242 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142257 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142276 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142292 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142307 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142324 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142341 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142355 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142372 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142390 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142407 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142425 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142440 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142456 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142471 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142488 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142504 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142520 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142544 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142563 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142578 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142628 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142644 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142660 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142679 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142695 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142713 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142732 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142750 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142819 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141805 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141774 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.141919 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142036 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142121 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142134 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142205 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142238 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.146223 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142464 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142492 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142500 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142664 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142684 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142839 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.142912 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.143136 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.143267 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.144594 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.144743 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.144886 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.145029 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.145157 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.145242 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.145278 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.145348 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.145497 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.145537 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.145780 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.145992 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.146195 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.146563 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.146596 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.146550 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.146766 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.146946 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.147037 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.147081 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.147123 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.147161 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.147166 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.147232 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.147290 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.147318 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.147538 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.147596 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.148014 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.148030 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.148147 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.148512 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.148554 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.148685 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.148648 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.148799 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.148848 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.148950 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.149054 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.149489 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.149530 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.149954 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.150470 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.151258 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.151288 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.151314 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.151316 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.151335 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.151375 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.151612 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.151665 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.151726 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.151923 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.152097 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.152318 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.152413 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.152432 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.152573 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.152691 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.152769 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.153957 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.153947 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.154060 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.154099 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.154093 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.154224 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.154533 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.154572 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.154782 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.154849 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.154859 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.154869 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.154876 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.152904 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155057 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155097 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155114 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155130 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155146 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155162 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155178 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155194 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155210 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155225 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155242 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155257 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155273 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155289 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155304 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155320 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155335 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155355 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155361 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155370 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155411 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155437 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155493 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155575 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155597 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.152215 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155624 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155660 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155666 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155672 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155690 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155777 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155824 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155845 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155888 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155912 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155935 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155958 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155979 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156001 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156022 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156044 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156066 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156089 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156137 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156162 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156280 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156306 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156331 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156355 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156395 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156421 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156446 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156486 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156511 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156534 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156556 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156579 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156652 4989 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156676 4989 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156690 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156702 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156713 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156726 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156737 4989 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156749 4989 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156760 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156772 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156783 4989 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156817 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156831 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156846 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156880 4989 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156897 4989 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156909 4989 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156919 4989 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156931 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156942 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156953 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156964 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156975 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156986 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156998 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157011 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157023 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157034 4989 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157045 4989 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157076 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157087 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157098 4989 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157110 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157120 4989 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157131 4989 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157142 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157154 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157167 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157178 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157191 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157202 4989 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157213 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157224 4989 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157235 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157246 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157256 4989 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157267 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157279 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157290 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157300 4989 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157312 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157332 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157347 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157358 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157369 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157381 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157392 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157404 4989 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157415 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157426 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157437 4989 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157448 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157459 4989 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157470 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157480 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157492 4989 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157503 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157533 4989 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157544 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157555 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157566 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157579 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157591 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157603 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157614 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157626 4989 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157638 4989 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157650 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157662 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157678 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157690 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157702 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157715 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157726 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157744 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157758 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157769 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157781 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157811 4989 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157823 4989 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157837 4989 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157850 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157861 4989 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157873 4989 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157884 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157896 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157908 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157919 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157931 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157944 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157956 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157973 4989 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157983 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157995 4989 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158007 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158018 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158030 4989 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158041 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158052 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158063 4989 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158075 4989 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158086 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158098 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158109 4989 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158120 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158132 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158146 4989 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155683 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.161060 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.155942 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156011 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156012 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156034 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156219 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156286 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156396 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156408 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.161778 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.161937 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.161971 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.162255 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156480 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156503 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156584 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156599 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156625 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156808 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156828 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.156932 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.162524 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157172 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157271 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157277 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157373 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157640 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157656 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.162608 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.157830 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158065 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158190 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.158231 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158276 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158555 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158620 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.162778 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158675 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.162842 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.158782 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.159007 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.159003 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.159029 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.159126 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.159683 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.159720 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.162903 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.159829 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.159879 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.160180 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.160588 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.160657 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.160683 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.160704 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:16:44.66068138 +0000 UTC m=+19.267128528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.163073 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.163077 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.162881 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.163259 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.163399 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.163425 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.163537 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.163522 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.163562 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.163766 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.162905 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.163975 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.163988 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.163004 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.164100 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:44.664059998 +0000 UTC m=+19.270507186 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.164273 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:44.664240994 +0000 UTC m=+19.270688172 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.164379 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.164486 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.164527 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.164588 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.164781 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.164813 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.164880 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.165065 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.165578 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.166017 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.171503 4989 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.178282 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.178640 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.179189 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.179350 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.179657 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.181963 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.181982 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.181994 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.182041 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:44.682025459 +0000 UTC m=+19.288472597 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.182040 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.182065 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.182079 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.182128 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:44.682110521 +0000 UTC m=+19.288557739 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.186209 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.194830 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.197251 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.206995 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258829 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258862 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258910 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258919 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258929 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258937 4989 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258946 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258955 4989 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258963 4989 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258971 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258980 4989 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258988 4989 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.258997 4989 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259005 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259013 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259023 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259036 4989 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259048 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259059 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259070 4989 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259080 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259072 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259105 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259088 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259197 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259207 4989 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259217 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259226 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259252 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259262 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259269 4989 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259278 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259286 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259296 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259305 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259328 4989 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259337 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259346 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259354 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259364 4989 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259373 4989 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259382 4989 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259404 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259414 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259423 4989 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259432 4989 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259441 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259451 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259461 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259485 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259494 4989 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259503 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259512 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259521 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259529 4989 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259538 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259562 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259571 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259580 4989 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259589 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259598 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259606 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259617 4989 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259641 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259650 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259658 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259668 4989 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259680 4989 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259689 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259713 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259721 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259729 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259738 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259747 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259757 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259768 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259805 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259814 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259824 4989 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259839 4989 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.259870 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.279115 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.289944 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.299773 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 13 17:16:44 crc kubenswrapper[4989]: W1213 17:16:44.310006 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-85a4989194305a87420590a3a96de77107201bbe432c61af87c0c8adf238917e WatchSource:0}: Error finding container 85a4989194305a87420590a3a96de77107201bbe432c61af87c0c8adf238917e: Status 404 returned error can't find the container with id 85a4989194305a87420590a3a96de77107201bbe432c61af87c0c8adf238917e Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.519605 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.520370 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.525836 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.529288 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.533164 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.545890 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.554640 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.564350 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.576583 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.585719 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.594381 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.606858 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.617498 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.624957 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.633337 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.641260 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.649775 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.659738 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.663173 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.663303 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:16:45.663265477 +0000 UTC m=+20.269712615 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.669410 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.764757 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.764822 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.764849 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:44 crc kubenswrapper[4989]: I1213 17:16:44.764888 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.764944 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.764966 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.764988 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.765016 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.765029 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.765031 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:45.765005818 +0000 UTC m=+20.371452966 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.765066 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:45.765052819 +0000 UTC m=+20.371499957 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.765075 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.765089 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.765094 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:45.76507942 +0000 UTC m=+20.371526548 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.765100 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:44 crc kubenswrapper[4989]: E1213 17:16:44.765145 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:45.765123011 +0000 UTC m=+20.371570149 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.113827 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2113445b8dda135049fea3654040d92305d94e3ebdf0ec065262d9278ada735b"} Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.115213 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3"} Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.115246 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31"} Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.115255 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"85a4989194305a87420590a3a96de77107201bbe432c61af87c0c8adf238917e"} Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.116909 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529"} Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.116946 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"71594d832e18b4ff85b3aa9630b3bf23d6232bdd9d6d34c32b1ba01ecceac2eb"} Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.132502 4989 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.140772 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.153971 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.170196 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.191252 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.201978 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.214348 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.225604 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.240044 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.253531 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.269633 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.294324 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.309239 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.321877 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.335592 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.349098 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.363144 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:45Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.672180 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.672360 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:16:47.672342365 +0000 UTC m=+22.278789513 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.773547 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.773595 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.773620 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:45 crc kubenswrapper[4989]: I1213 17:16:45.773639 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.773730 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.773765 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.773807 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.773818 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.773836 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:47.773814728 +0000 UTC m=+22.380261866 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.773882 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:47.77386743 +0000 UTC m=+22.380314568 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.773741 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.773922 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.773930 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.773953 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:47.773947162 +0000 UTC m=+22.380394300 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.774058 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:45 crc kubenswrapper[4989]: E1213 17:16:45.774084 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:47.774077335 +0000 UTC m=+22.380524473 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.015003 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.015033 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:46 crc kubenswrapper[4989]: E1213 17:16:46.015127 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:16:46 crc kubenswrapper[4989]: E1213 17:16:46.015298 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.015334 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:46 crc kubenswrapper[4989]: E1213 17:16:46.015515 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.020850 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.021923 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.024373 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.025443 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.026238 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.026988 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.027936 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.028661 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.029509 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.030228 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.031111 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.031973 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.032494 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.033644 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.034113 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.034352 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.035242 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.035846 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.036245 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.037314 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.037887 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.038366 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.039264 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.039678 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.040633 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.041021 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.041985 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.042560 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.043352 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.043958 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.044737 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.045181 4989 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.045273 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.046775 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.047713 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.048102 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.049588 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.050571 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.050677 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.051299 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.052563 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.054245 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.056367 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.057771 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.060107 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.061646 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.063142 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.063812 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.064536 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.065392 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.065961 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.066492 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.068757 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.069318 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.070023 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.071954 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.072650 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.091319 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.106572 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.121143 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.132666 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:46 crc kubenswrapper[4989]: I1213 17:16:46.145563 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.038543 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.040306 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.040342 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.040356 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.040416 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.046441 4989 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.046917 4989 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.048400 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.048433 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.048444 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.048465 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.048480 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.072849 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.076222 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.076254 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.076264 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.076277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.076287 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.088334 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.091176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.091209 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.091220 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.091236 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.091246 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.101875 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.104898 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.104934 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.104945 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.104959 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.104971 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.115408 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.118445 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.118490 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.118504 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.118525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.118541 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.121167 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793"} Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.130508 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.130657 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.132155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.132197 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.132212 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.132229 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.132240 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.133434 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.144023 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.153311 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.162631 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.172364 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.183419 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.196164 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.207623 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:47Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.234404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.234430 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.234438 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.234449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.234458 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.336458 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.336497 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.336510 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.336526 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.336537 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.439349 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.439404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.439417 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.439435 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.439448 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.542657 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.542690 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.542702 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.542718 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.542729 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.644673 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.644714 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.644724 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.644736 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.644745 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.687269 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.687466 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:16:51.687440701 +0000 UTC m=+26.293887839 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.748074 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.748158 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.748190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.748229 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.748255 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.788735 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.788867 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.788936 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.788994 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789095 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789151 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789161 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789183 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789193 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789182 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789216 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789290 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:51.789256174 +0000 UTC m=+26.395703352 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789161 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789323 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:51.789309486 +0000 UTC m=+26.395756664 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789420 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:51.789389528 +0000 UTC m=+26.395836716 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:47 crc kubenswrapper[4989]: E1213 17:16:47.789493 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:51.789469881 +0000 UTC m=+26.395917069 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.851645 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.851708 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.851735 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.851763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.851780 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.955029 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.955076 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.955085 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.955102 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:47 crc kubenswrapper[4989]: I1213 17:16:47.955113 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:47Z","lastTransitionTime":"2025-12-13T17:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.014432 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.014485 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.014461 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:48 crc kubenswrapper[4989]: E1213 17:16:48.014715 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:16:48 crc kubenswrapper[4989]: E1213 17:16:48.014820 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:16:48 crc kubenswrapper[4989]: E1213 17:16:48.014895 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.058041 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.058084 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.058095 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.058112 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.058122 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:48Z","lastTransitionTime":"2025-12-13T17:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.160201 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.160247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.160259 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.160273 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.160284 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:48Z","lastTransitionTime":"2025-12-13T17:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.263010 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.263059 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.263069 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.263084 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.263094 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:48Z","lastTransitionTime":"2025-12-13T17:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.365951 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.366019 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.366032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.366057 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.366073 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:48Z","lastTransitionTime":"2025-12-13T17:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.469339 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.469389 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.469399 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.469417 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.469428 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:48Z","lastTransitionTime":"2025-12-13T17:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.572230 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.572310 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.572330 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.572363 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.572384 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:48Z","lastTransitionTime":"2025-12-13T17:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.675542 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.675614 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.675637 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.675669 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.675689 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:48Z","lastTransitionTime":"2025-12-13T17:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.779629 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.779714 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.779731 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.779758 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.779776 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:48Z","lastTransitionTime":"2025-12-13T17:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.881948 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.882017 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.882035 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.882060 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.882079 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:48Z","lastTransitionTime":"2025-12-13T17:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.919963 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.936440 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.937552 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.937876 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:48Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.952890 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:48Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.966130 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:48Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.983310 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:48Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.984891 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.984940 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.984958 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.984974 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.984987 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:48Z","lastTransitionTime":"2025-12-13T17:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:48 crc kubenswrapper[4989]: I1213 17:16:48.998945 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:48Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.016439 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.035141 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.049935 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.076618 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.088026 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.088067 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.088080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.088120 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.088132 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:49Z","lastTransitionTime":"2025-12-13T17:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.096014 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.110439 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.129280 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: E1213 17:16:49.138825 4989 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.141602 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.155033 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.174325 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.190761 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.190805 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.190815 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.190829 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.190839 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:49Z","lastTransitionTime":"2025-12-13T17:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.195911 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.210285 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:49Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.293979 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.294046 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.294066 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.294093 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.294112 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:49Z","lastTransitionTime":"2025-12-13T17:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.398251 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.398314 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.398327 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.398351 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.398365 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:49Z","lastTransitionTime":"2025-12-13T17:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.502092 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.502140 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.502157 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.502178 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.502192 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:49Z","lastTransitionTime":"2025-12-13T17:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.604710 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.604754 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.604764 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.604778 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.604807 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:49Z","lastTransitionTime":"2025-12-13T17:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.706725 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.706766 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.706777 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.706807 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.706818 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:49Z","lastTransitionTime":"2025-12-13T17:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.808831 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.808875 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.808884 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.808898 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.808908 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:49Z","lastTransitionTime":"2025-12-13T17:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.911631 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.911663 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.911672 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.911685 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:49 crc kubenswrapper[4989]: I1213 17:16:49.911694 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:49Z","lastTransitionTime":"2025-12-13T17:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.013593 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.013727 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:50 crc kubenswrapper[4989]: E1213 17:16:50.013863 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:16:50 crc kubenswrapper[4989]: E1213 17:16:50.013935 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.014062 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.014410 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.014459 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.014481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.014508 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.014534 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:50Z","lastTransitionTime":"2025-12-13T17:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:50 crc kubenswrapper[4989]: E1213 17:16:50.014423 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.116898 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.116933 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.116942 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.116956 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.116965 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:50Z","lastTransitionTime":"2025-12-13T17:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.219725 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.220321 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.220401 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.220482 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.220546 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:50Z","lastTransitionTime":"2025-12-13T17:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.322856 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.322930 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.322949 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.322975 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.322994 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:50Z","lastTransitionTime":"2025-12-13T17:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.425330 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.425378 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.425393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.425414 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.425445 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:50Z","lastTransitionTime":"2025-12-13T17:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.528231 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.528262 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.528271 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.528285 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.528296 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:50Z","lastTransitionTime":"2025-12-13T17:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.630910 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.630966 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.630977 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.631001 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.631015 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:50Z","lastTransitionTime":"2025-12-13T17:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.733261 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.733321 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.733332 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.733349 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.733359 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:50Z","lastTransitionTime":"2025-12-13T17:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.836151 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.836184 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.836193 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.836206 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.836214 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:50Z","lastTransitionTime":"2025-12-13T17:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.856316 4989 csr.go:261] certificate signing request csr-qzh49 is approved, waiting to be issued Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.865989 4989 csr.go:257] certificate signing request csr-qzh49 is issued Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.939174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.939416 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.939488 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.939561 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:50 crc kubenswrapper[4989]: I1213 17:16:50.939617 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:50Z","lastTransitionTime":"2025-12-13T17:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.041357 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.041592 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.041724 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.041842 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.041930 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:51Z","lastTransitionTime":"2025-12-13T17:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.143657 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.143698 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.143707 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.143720 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.143732 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:51Z","lastTransitionTime":"2025-12-13T17:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.246365 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.246907 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.247015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.247127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.247191 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:51Z","lastTransitionTime":"2025-12-13T17:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.349618 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.349841 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.349929 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.350005 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.350061 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:51Z","lastTransitionTime":"2025-12-13T17:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.452137 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.452167 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.452175 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.452189 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.452199 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:51Z","lastTransitionTime":"2025-12-13T17:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.554154 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.554195 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.554207 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.554222 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.554232 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:51Z","lastTransitionTime":"2025-12-13T17:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.590642 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-hllvq"] Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.590952 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.592470 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-dv9sz"] Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.593187 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.596173 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-nh9k2"] Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.596559 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.604195 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.604318 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-n4l98"] Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.604475 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.604649 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-n4l98" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.606553 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 13 17:16:51 crc kubenswrapper[4989]: W1213 17:16:51.607849 4989 reflector.go:561] object-"openshift-machine-config-operator"/"proxy-tls": failed to list *v1.Secret: secrets "proxy-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.607879 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.607893 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"proxy-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"proxy-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:16:51 crc kubenswrapper[4989]: W1213 17:16:51.608430 4989 reflector.go:561] object-"openshift-machine-config-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.608461 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:16:51 crc kubenswrapper[4989]: W1213 17:16:51.608520 4989 reflector.go:561] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.608532 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.609110 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 13 17:16:51 crc kubenswrapper[4989]: W1213 17:16:51.609128 4989 reflector.go:561] object-"openshift-machine-config-operator"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.609146 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.609156 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 13 17:16:51 crc kubenswrapper[4989]: W1213 17:16:51.609180 4989 reflector.go:561] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": failed to list *v1.Secret: secrets "node-resolver-dockercfg-kz9s7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 13 17:16:51 crc kubenswrapper[4989]: W1213 17:16:51.609190 4989 reflector.go:561] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": failed to list *v1.Secret: secrets "machine-config-daemon-dockercfg-r5tcq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.609206 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-r5tcq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-config-daemon-dockercfg-r5tcq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.609220 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"node-resolver-dockercfg-kz9s7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"node-resolver-dockercfg-kz9s7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.609446 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.611260 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.611349 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.619939 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/59ed6e54-65fe-4383-9578-d0c89a69ecec-cni-binary-copy\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.619986 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-system-cni-dir\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620013 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a2b01148-171a-4f86-84a7-d326739e0dcf-rootfs\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620031 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-run-netns\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620046 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-var-lib-cni-multus\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620062 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-system-cni-dir\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620076 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-var-lib-kubelet\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620090 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-run-multus-certs\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620103 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-cnibin\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620118 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/59ed6e54-65fe-4383-9578-d0c89a69ecec-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620141 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-socket-dir-parent\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620154 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a2b01148-171a-4f86-84a7-d326739e0dcf-proxy-tls\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620170 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a2b01148-171a-4f86-84a7-d326739e0dcf-mcd-auth-proxy-config\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620185 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-cnibin\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620200 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/851885b5-3fdc-4e01-87d2-4a79a73acd6a-hosts-file\") pod \"node-resolver-n4l98\" (UID: \"851885b5-3fdc-4e01-87d2-4a79a73acd6a\") " pod="openshift-dns/node-resolver-n4l98" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620213 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-hostroot\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620229 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hpvp\" (UniqueName: \"kubernetes.io/projected/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-kube-api-access-6hpvp\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620243 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-os-release\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620256 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdsjc\" (UniqueName: \"kubernetes.io/projected/59ed6e54-65fe-4383-9578-d0c89a69ecec-kube-api-access-cdsjc\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620271 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-conf-dir\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620286 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-daemon-config\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620299 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcv48\" (UniqueName: \"kubernetes.io/projected/851885b5-3fdc-4e01-87d2-4a79a73acd6a-kube-api-access-vcv48\") pod \"node-resolver-n4l98\" (UID: \"851885b5-3fdc-4e01-87d2-4a79a73acd6a\") " pod="openshift-dns/node-resolver-n4l98" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620315 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-os-release\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620329 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-run-k8s-cni-cncf-io\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620343 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-etc-kubernetes\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620358 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn7p6\" (UniqueName: \"kubernetes.io/projected/a2b01148-171a-4f86-84a7-d326739e0dcf-kube-api-access-dn7p6\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620384 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-cni-binary-copy\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620397 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-cni-dir\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620409 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-var-lib-cni-bin\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.620426 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.638083 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.656253 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.656302 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.656317 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.656336 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.656348 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:51Z","lastTransitionTime":"2025-12-13T17:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.664685 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.679701 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.694058 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.707203 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.718179 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.720894 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.721046 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:16:59.721024762 +0000 UTC m=+34.327471910 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721089 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-conf-dir\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721120 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-daemon-config\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721138 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcv48\" (UniqueName: \"kubernetes.io/projected/851885b5-3fdc-4e01-87d2-4a79a73acd6a-kube-api-access-vcv48\") pod \"node-resolver-n4l98\" (UID: \"851885b5-3fdc-4e01-87d2-4a79a73acd6a\") " pod="openshift-dns/node-resolver-n4l98" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721157 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-os-release\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721173 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-run-k8s-cni-cncf-io\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721187 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-etc-kubernetes\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721203 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn7p6\" (UniqueName: \"kubernetes.io/projected/a2b01148-171a-4f86-84a7-d326739e0dcf-kube-api-access-dn7p6\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721235 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-cni-binary-copy\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721250 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-cni-dir\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721264 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-var-lib-cni-bin\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721279 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721305 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/59ed6e54-65fe-4383-9578-d0c89a69ecec-cni-binary-copy\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721321 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-system-cni-dir\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721337 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a2b01148-171a-4f86-84a7-d326739e0dcf-rootfs\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721342 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-conf-dir\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721352 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-run-netns\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721383 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-run-netns\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721397 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-var-lib-cni-multus\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721419 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-system-cni-dir\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721411 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-run-k8s-cni-cncf-io\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721434 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-var-lib-kubelet\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721447 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-etc-kubernetes\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721451 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-run-multus-certs\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721469 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-cnibin\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721485 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/59ed6e54-65fe-4383-9578-d0c89a69ecec-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721513 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-socket-dir-parent\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721527 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a2b01148-171a-4f86-84a7-d326739e0dcf-proxy-tls\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721543 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a2b01148-171a-4f86-84a7-d326739e0dcf-mcd-auth-proxy-config\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721558 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-cnibin\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721574 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/851885b5-3fdc-4e01-87d2-4a79a73acd6a-hosts-file\") pod \"node-resolver-n4l98\" (UID: \"851885b5-3fdc-4e01-87d2-4a79a73acd6a\") " pod="openshift-dns/node-resolver-n4l98" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721588 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-hostroot\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721603 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hpvp\" (UniqueName: \"kubernetes.io/projected/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-kube-api-access-6hpvp\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721629 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-os-release\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721646 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdsjc\" (UniqueName: \"kubernetes.io/projected/59ed6e54-65fe-4383-9578-d0c89a69ecec-kube-api-access-cdsjc\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.722176 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-cni-dir\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.722288 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-socket-dir-parent\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.722325 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-var-lib-cni-multus\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.722370 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-system-cni-dir\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.722407 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-var-lib-kubelet\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.722437 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-cnibin\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.722441 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-run-multus-certs\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.722467 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-cnibin\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.722509 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/851885b5-3fdc-4e01-87d2-4a79a73acd6a-hosts-file\") pod \"node-resolver-n4l98\" (UID: \"851885b5-3fdc-4e01-87d2-4a79a73acd6a\") " pod="openshift-dns/node-resolver-n4l98" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.722556 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-hostroot\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.722730 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-os-release\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.723095 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-system-cni-dir\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.721317 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-os-release\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.723240 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-host-var-lib-cni-bin\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.723216 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/59ed6e54-65fe-4383-9578-d0c89a69ecec-cni-binary-copy\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.723263 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/59ed6e54-65fe-4383-9578-d0c89a69ecec-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.723381 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a2b01148-171a-4f86-84a7-d326739e0dcf-rootfs\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.723681 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-cni-binary-copy\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.723806 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-multus-daemon-config\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.723845 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/59ed6e54-65fe-4383-9578-d0c89a69ecec-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.736117 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.737613 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdsjc\" (UniqueName: \"kubernetes.io/projected/59ed6e54-65fe-4383-9578-d0c89a69ecec-kube-api-access-cdsjc\") pod \"multus-additional-cni-plugins-dv9sz\" (UID: \"59ed6e54-65fe-4383-9578-d0c89a69ecec\") " pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.738213 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcv48\" (UniqueName: \"kubernetes.io/projected/851885b5-3fdc-4e01-87d2-4a79a73acd6a-kube-api-access-vcv48\") pod \"node-resolver-n4l98\" (UID: \"851885b5-3fdc-4e01-87d2-4a79a73acd6a\") " pod="openshift-dns/node-resolver-n4l98" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.739587 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hpvp\" (UniqueName: \"kubernetes.io/projected/1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf-kube-api-access-6hpvp\") pod \"multus-hllvq\" (UID: \"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\") " pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.748664 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.760923 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.761023 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.761037 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.761075 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.761090 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:51Z","lastTransitionTime":"2025-12-13T17:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.761414 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.773227 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.785251 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.797946 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.815192 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.822402 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.822537 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.822743 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.822640 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.822903 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.822920 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.822975 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:59.822958409 +0000 UTC m=+34.429405547 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.822761 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.822866 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.822860 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.823282 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:59.823203086 +0000 UTC m=+34.429650224 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.823322 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.823382 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.823457 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:59.823437263 +0000 UTC m=+34.429884401 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.823545 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:51 crc kubenswrapper[4989]: E1213 17:16:51.823632 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:16:59.823623939 +0000 UTC m=+34.430071077 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.829180 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.837224 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.854130 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.863409 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.863455 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.863468 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.863486 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.863498 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:51Z","lastTransitionTime":"2025-12-13T17:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.865671 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.867680 4989 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-13 17:11:50 +0000 UTC, rotation deadline is 2026-09-01 20:38:16.402170194 +0000 UTC Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.867734 4989 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6291h21m24.534439026s for next certificate rotation Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.877080 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.886711 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.897754 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.902851 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hllvq" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.908849 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.912195 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: W1213 17:16:51.915270 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ea718eb_ab21_4f3c_8d0d_c6cf4ffe69bf.slice/crio-e45729dbee71a179c7c5026510eb1e767315a25e5f8516ed4e219f9aa9a28de2 WatchSource:0}: Error finding container e45729dbee71a179c7c5026510eb1e767315a25e5f8516ed4e219f9aa9a28de2: Status 404 returned error can't find the container with id e45729dbee71a179c7c5026510eb1e767315a25e5f8516ed4e219f9aa9a28de2 Dec 13 17:16:51 crc kubenswrapper[4989]: W1213 17:16:51.922398 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59ed6e54_65fe_4383_9578_d0c89a69ecec.slice/crio-d84a46979c39e8b521f77a5699b7bb61b2b180626c1ccb685bc2a733a62f435f WatchSource:0}: Error finding container d84a46979c39e8b521f77a5699b7bb61b2b180626c1ccb685bc2a733a62f435f: Status 404 returned error can't find the container with id d84a46979c39e8b521f77a5699b7bb61b2b180626c1ccb685bc2a733a62f435f Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.926861 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.942084 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.966487 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.966520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.966530 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.966744 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.966756 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:51Z","lastTransitionTime":"2025-12-13T17:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.974997 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-z7249"] Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.975890 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.978910 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.979361 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.979587 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.979695 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.979822 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.980004 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.980215 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 13 17:16:51 crc kubenswrapper[4989]: I1213 17:16:51.993600 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:51Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.007045 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.015781 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:52 crc kubenswrapper[4989]: E1213 17:16:52.015940 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.015994 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.016023 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:52 crc kubenswrapper[4989]: E1213 17:16:52.016107 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:16:52 crc kubenswrapper[4989]: E1213 17:16:52.016216 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.021554 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024372 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-var-lib-openvswitch\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024410 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/101724b9-153f-4f9d-849a-c04a343e7446-ovn-node-metrics-cert\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024436 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blhj7\" (UniqueName: \"kubernetes.io/projected/101724b9-153f-4f9d-849a-c04a343e7446-kube-api-access-blhj7\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024473 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-openvswitch\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024490 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-bin\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024507 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-ovn-kubernetes\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024522 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-slash\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024551 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-kubelet\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024566 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-systemd\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024582 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-netd\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024597 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-node-log\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024612 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-ovn\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024628 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-etc-openvswitch\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024644 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-netns\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024660 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024679 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-env-overrides\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024704 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-systemd-units\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024720 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-log-socket\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024737 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-config\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.024753 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-script-lib\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.036652 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.047747 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.069545 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.078529 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.078580 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.078594 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.078617 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.078632 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:52Z","lastTransitionTime":"2025-12-13T17:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.082698 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.094679 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.105350 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127246 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-script-lib\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127305 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-var-lib-openvswitch\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127331 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/101724b9-153f-4f9d-849a-c04a343e7446-ovn-node-metrics-cert\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127368 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blhj7\" (UniqueName: \"kubernetes.io/projected/101724b9-153f-4f9d-849a-c04a343e7446-kube-api-access-blhj7\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127407 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-bin\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127437 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-openvswitch\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127458 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-slash\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127390 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127522 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-ovn-kubernetes\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127570 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-var-lib-openvswitch\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127479 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-ovn-kubernetes\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127664 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-kubelet\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127682 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-systemd\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127710 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-node-log\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127730 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-netd\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127754 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-ovn\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127775 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-etc-openvswitch\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127809 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127826 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-env-overrides\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127849 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-netns\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127876 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-systemd-units\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127891 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-log-socket\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.127906 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-config\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128152 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-script-lib\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128203 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-bin\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128214 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-openvswitch\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128265 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-etc-openvswitch\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128283 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-kubelet\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128314 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-systemd\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128317 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-slash\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128331 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-netns\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128365 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-systemd-units\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128346 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-ovn\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128361 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128396 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-log-socket\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128399 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-node-log\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128466 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-netd\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.128891 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-config\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.129297 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-env-overrides\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.134248 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/101724b9-153f-4f9d-849a-c04a343e7446-ovn-node-metrics-cert\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.137483 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" event={"ID":"59ed6e54-65fe-4383-9578-d0c89a69ecec","Type":"ContainerStarted","Data":"d84a46979c39e8b521f77a5699b7bb61b2b180626c1ccb685bc2a733a62f435f"} Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.139549 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.140654 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hllvq" event={"ID":"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf","Type":"ContainerStarted","Data":"d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494"} Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.140691 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hllvq" event={"ID":"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf","Type":"ContainerStarted","Data":"e45729dbee71a179c7c5026510eb1e767315a25e5f8516ed4e219f9aa9a28de2"} Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.144741 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blhj7\" (UniqueName: \"kubernetes.io/projected/101724b9-153f-4f9d-849a-c04a343e7446-kube-api-access-blhj7\") pod \"ovnkube-node-z7249\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.151050 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.163184 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.175439 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.180932 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.180971 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.180980 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.180993 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.181003 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:52Z","lastTransitionTime":"2025-12-13T17:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.186406 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.200722 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.212875 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.226570 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.241270 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.253927 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.267681 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.278775 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.283356 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.283384 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.283393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.283406 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.283447 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:52Z","lastTransitionTime":"2025-12-13T17:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.288023 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.292618 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: W1213 17:16:52.299767 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod101724b9_153f_4f9d_849a_c04a343e7446.slice/crio-bafb9bdcc89878b80e073d2652fbd3ea185f04e42ad8f704985f0187ceea4b27 WatchSource:0}: Error finding container bafb9bdcc89878b80e073d2652fbd3ea185f04e42ad8f704985f0187ceea4b27: Status 404 returned error can't find the container with id bafb9bdcc89878b80e073d2652fbd3ea185f04e42ad8f704985f0187ceea4b27 Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.306342 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.315876 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.331935 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.351856 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.364922 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:52Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.385465 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.385493 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.385501 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.385513 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.385523 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:52Z","lastTransitionTime":"2025-12-13T17:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.487869 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.487896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.487918 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.487932 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.487941 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:52Z","lastTransitionTime":"2025-12-13T17:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.548245 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.590692 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.590726 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.590743 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.590756 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.590765 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:52Z","lastTransitionTime":"2025-12-13T17:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.671126 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.692964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.693000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.693007 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.693022 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.693031 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:52Z","lastTransitionTime":"2025-12-13T17:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:52 crc kubenswrapper[4989]: E1213 17:16:52.722843 4989 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:16:52 crc kubenswrapper[4989]: E1213 17:16:52.722853 4989 secret.go:188] Couldn't get secret openshift-machine-config-operator/proxy-tls: failed to sync secret cache: timed out waiting for the condition Dec 13 17:16:52 crc kubenswrapper[4989]: E1213 17:16:52.722924 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a2b01148-171a-4f86-84a7-d326739e0dcf-mcd-auth-proxy-config podName:a2b01148-171a-4f86-84a7-d326739e0dcf nodeName:}" failed. No retries permitted until 2025-12-13 17:16:53.222906752 +0000 UTC m=+27.829353890 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "mcd-auth-proxy-config" (UniqueName: "kubernetes.io/configmap/a2b01148-171a-4f86-84a7-d326739e0dcf-mcd-auth-proxy-config") pod "machine-config-daemon-nh9k2" (UID: "a2b01148-171a-4f86-84a7-d326739e0dcf") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:16:52 crc kubenswrapper[4989]: E1213 17:16:52.722943 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a2b01148-171a-4f86-84a7-d326739e0dcf-proxy-tls podName:a2b01148-171a-4f86-84a7-d326739e0dcf nodeName:}" failed. No retries permitted until 2025-12-13 17:16:53.222935203 +0000 UTC m=+27.829382341 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/a2b01148-171a-4f86-84a7-d326739e0dcf-proxy-tls") pod "machine-config-daemon-nh9k2" (UID: "a2b01148-171a-4f86-84a7-d326739e0dcf") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.795709 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.795747 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.795756 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.795771 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.795781 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:52Z","lastTransitionTime":"2025-12-13T17:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.836220 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.841292 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-n4l98" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.843041 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.897682 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.897719 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.897730 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.897745 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:52 crc kubenswrapper[4989]: I1213 17:16:52.897758 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:52Z","lastTransitionTime":"2025-12-13T17:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.001644 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.001692 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.001706 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.001727 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.001740 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:53Z","lastTransitionTime":"2025-12-13T17:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.010229 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.017525 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn7p6\" (UniqueName: \"kubernetes.io/projected/a2b01148-171a-4f86-84a7-d326739e0dcf-kube-api-access-dn7p6\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.106051 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.106102 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.106117 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.106138 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.106151 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:53Z","lastTransitionTime":"2025-12-13T17:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.108242 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.145684 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671" exitCode=0 Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.145736 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.145880 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"bafb9bdcc89878b80e073d2652fbd3ea185f04e42ad8f704985f0187ceea4b27"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.148343 4989 generic.go:334] "Generic (PLEG): container finished" podID="59ed6e54-65fe-4383-9578-d0c89a69ecec" containerID="0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6" exitCode=0 Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.148405 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" event={"ID":"59ed6e54-65fe-4383-9578-d0c89a69ecec","Type":"ContainerDied","Data":"0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.150478 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-n4l98" event={"ID":"851885b5-3fdc-4e01-87d2-4a79a73acd6a","Type":"ContainerStarted","Data":"8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.150513 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-n4l98" event={"ID":"851885b5-3fdc-4e01-87d2-4a79a73acd6a","Type":"ContainerStarted","Data":"34e29c32d448890bc00a8ad5d6c061cec4bc8b5aea5479b24c11da9abace149e"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.164004 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.179776 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.199744 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.209935 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.209965 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.209973 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.209987 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.209998 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:53Z","lastTransitionTime":"2025-12-13T17:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.215160 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.230087 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.241313 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a2b01148-171a-4f86-84a7-d326739e0dcf-proxy-tls\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.241370 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a2b01148-171a-4f86-84a7-d326739e0dcf-mcd-auth-proxy-config\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.249255 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.249650 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a2b01148-171a-4f86-84a7-d326739e0dcf-mcd-auth-proxy-config\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.250723 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a2b01148-171a-4f86-84a7-d326739e0dcf-proxy-tls\") pod \"machine-config-daemon-nh9k2\" (UID: \"a2b01148-171a-4f86-84a7-d326739e0dcf\") " pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.289026 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.322101 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.322356 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.322414 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.322429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.322451 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.322464 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:53Z","lastTransitionTime":"2025-12-13T17:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.367044 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.389819 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.415454 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.416551 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.430902 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.430976 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.430987 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.431006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.431066 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:53Z","lastTransitionTime":"2025-12-13T17:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.434400 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.447970 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.466610 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.480206 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.495439 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.512967 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.529054 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.536399 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.536449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.536460 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.536478 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.536488 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:53Z","lastTransitionTime":"2025-12-13T17:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.545161 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.569250 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.582220 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.597821 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.612127 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.630639 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.642900 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.642951 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.642964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.642983 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.642998 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:53Z","lastTransitionTime":"2025-12-13T17:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.665406 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.687951 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.707236 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.723654 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:53Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.745481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.745512 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.745520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.745538 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.745547 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:53Z","lastTransitionTime":"2025-12-13T17:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.847815 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.847880 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.847892 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.847931 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.847943 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:53Z","lastTransitionTime":"2025-12-13T17:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.950768 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.950826 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.950837 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.950853 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:53 crc kubenswrapper[4989]: I1213 17:16:53.950864 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:53Z","lastTransitionTime":"2025-12-13T17:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.014073 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.014136 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.014141 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:54 crc kubenswrapper[4989]: E1213 17:16:54.014216 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:16:54 crc kubenswrapper[4989]: E1213 17:16:54.014390 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:16:54 crc kubenswrapper[4989]: E1213 17:16:54.014658 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.053695 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.053740 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.053751 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.053767 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.053778 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:54Z","lastTransitionTime":"2025-12-13T17:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.158437 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.158469 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.158477 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.158491 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.158501 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:54Z","lastTransitionTime":"2025-12-13T17:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.160012 4989 generic.go:334] "Generic (PLEG): container finished" podID="59ed6e54-65fe-4383-9578-d0c89a69ecec" containerID="ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565" exitCode=0 Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.160172 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" event={"ID":"59ed6e54-65fe-4383-9578-d0c89a69ecec","Type":"ContainerDied","Data":"ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.166634 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerStarted","Data":"e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.166851 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerStarted","Data":"8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.166944 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerStarted","Data":"68f69b6f062959e47672d5685a687b5bb6a651522e88b3f00925b4b9848c265d"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.176815 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.177128 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.177163 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.177177 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.177189 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.177199 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.177209 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.189967 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.203323 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.216573 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.234511 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.250169 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.265121 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.265158 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.265167 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.265181 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.265192 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:54Z","lastTransitionTime":"2025-12-13T17:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.270821 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.282259 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.303044 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.315486 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.326736 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.335831 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.355020 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.367321 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.367825 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.367863 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.367874 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.367890 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.367903 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:54Z","lastTransitionTime":"2025-12-13T17:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.378124 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.388937 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.400658 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.412943 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.422687 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.432305 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.445332 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.454661 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.470190 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.470868 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.470981 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.471069 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.471197 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.471284 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:54Z","lastTransitionTime":"2025-12-13T17:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.480701 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.492315 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.501929 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.517892 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.540732 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:54Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.573337 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.573364 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.573371 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.573383 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.573400 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:54Z","lastTransitionTime":"2025-12-13T17:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.677213 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.677285 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.677307 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.677340 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.677363 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:54Z","lastTransitionTime":"2025-12-13T17:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.781236 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.781539 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.781554 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.781570 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.781581 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:54Z","lastTransitionTime":"2025-12-13T17:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.885622 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.885678 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.885694 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.885717 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.885734 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:54Z","lastTransitionTime":"2025-12-13T17:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.988412 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.988482 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.988492 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.988527 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:54 crc kubenswrapper[4989]: I1213 17:16:54.988539 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:54Z","lastTransitionTime":"2025-12-13T17:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.083502 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.090913 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.090940 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.090948 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.090976 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.090984 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:55Z","lastTransitionTime":"2025-12-13T17:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.100784 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.115377 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.131565 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.145282 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.168882 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.182570 4989 generic.go:334] "Generic (PLEG): container finished" podID="59ed6e54-65fe-4383-9578-d0c89a69ecec" containerID="8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f" exitCode=0 Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.183025 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" event={"ID":"59ed6e54-65fe-4383-9578-d0c89a69ecec","Type":"ContainerDied","Data":"8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f"} Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.192734 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.193628 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.193665 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.193680 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.193700 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.193718 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:55Z","lastTransitionTime":"2025-12-13T17:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.216021 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.249181 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.281108 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.294644 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.299679 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-vmx98"] Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.300391 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vmx98" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.302476 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.302507 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.302519 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.302581 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.302595 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:55Z","lastTransitionTime":"2025-12-13T17:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.305246 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.305751 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.306205 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.306511 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.315598 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.331715 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.347547 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.364483 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.390809 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25dd508e-594b-4f32-af84-61ee8d65f38b-host\") pod \"node-ca-vmx98\" (UID: \"25dd508e-594b-4f32-af84-61ee8d65f38b\") " pod="openshift-image-registry/node-ca-vmx98" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.391408 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/25dd508e-594b-4f32-af84-61ee8d65f38b-serviceca\") pod \"node-ca-vmx98\" (UID: \"25dd508e-594b-4f32-af84-61ee8d65f38b\") " pod="openshift-image-registry/node-ca-vmx98" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.392061 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.392264 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24j5k\" (UniqueName: \"kubernetes.io/projected/25dd508e-594b-4f32-af84-61ee8d65f38b-kube-api-access-24j5k\") pod \"node-ca-vmx98\" (UID: \"25dd508e-594b-4f32-af84-61ee8d65f38b\") " pod="openshift-image-registry/node-ca-vmx98" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.406538 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.407629 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.407653 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.407662 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.407675 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.407686 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:55Z","lastTransitionTime":"2025-12-13T17:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.421268 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.435624 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.457421 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.470115 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.483126 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.493826 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24j5k\" (UniqueName: \"kubernetes.io/projected/25dd508e-594b-4f32-af84-61ee8d65f38b-kube-api-access-24j5k\") pod \"node-ca-vmx98\" (UID: \"25dd508e-594b-4f32-af84-61ee8d65f38b\") " pod="openshift-image-registry/node-ca-vmx98" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.493918 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25dd508e-594b-4f32-af84-61ee8d65f38b-host\") pod \"node-ca-vmx98\" (UID: \"25dd508e-594b-4f32-af84-61ee8d65f38b\") " pod="openshift-image-registry/node-ca-vmx98" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.493958 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/25dd508e-594b-4f32-af84-61ee8d65f38b-serviceca\") pod \"node-ca-vmx98\" (UID: \"25dd508e-594b-4f32-af84-61ee8d65f38b\") " pod="openshift-image-registry/node-ca-vmx98" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.497187 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/25dd508e-594b-4f32-af84-61ee8d65f38b-host\") pod \"node-ca-vmx98\" (UID: \"25dd508e-594b-4f32-af84-61ee8d65f38b\") " pod="openshift-image-registry/node-ca-vmx98" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.498181 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/25dd508e-594b-4f32-af84-61ee8d65f38b-serviceca\") pod \"node-ca-vmx98\" (UID: \"25dd508e-594b-4f32-af84-61ee8d65f38b\") " pod="openshift-image-registry/node-ca-vmx98" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.500803 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.509700 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.509741 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.509753 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.509772 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.509814 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:55Z","lastTransitionTime":"2025-12-13T17:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.529515 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24j5k\" (UniqueName: \"kubernetes.io/projected/25dd508e-594b-4f32-af84-61ee8d65f38b-kube-api-access-24j5k\") pod \"node-ca-vmx98\" (UID: \"25dd508e-594b-4f32-af84-61ee8d65f38b\") " pod="openshift-image-registry/node-ca-vmx98" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.561151 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.599656 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.613185 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.613225 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.613236 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.613252 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.613262 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:55Z","lastTransitionTime":"2025-12-13T17:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.623630 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vmx98" Dec 13 17:16:55 crc kubenswrapper[4989]: W1213 17:16:55.640316 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25dd508e_594b_4f32_af84_61ee8d65f38b.slice/crio-c95c37c96164687c39042035e0dc90effc36017181ba851b19d91e8ccec0937e WatchSource:0}: Error finding container c95c37c96164687c39042035e0dc90effc36017181ba851b19d91e8ccec0937e: Status 404 returned error can't find the container with id c95c37c96164687c39042035e0dc90effc36017181ba851b19d91e8ccec0937e Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.645616 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.687950 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.714907 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.714935 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.714945 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.714958 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.714967 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:55Z","lastTransitionTime":"2025-12-13T17:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.720780 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.765053 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.801921 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:55Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.817661 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.817709 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.817719 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.817732 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.817741 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:55Z","lastTransitionTime":"2025-12-13T17:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.857635 4989 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 13 17:16:55 crc kubenswrapper[4989]: W1213 17:16:55.859045 4989 reflector.go:484] object-"openshift-image-registry"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 13 17:16:55 crc kubenswrapper[4989]: W1213 17:16:55.859286 4989 reflector.go:484] object-"openshift-image-registry"/"image-registry-certificates": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"image-registry-certificates": Unexpected watch close - watch lasted less than a second and no items received Dec 13 17:16:55 crc kubenswrapper[4989]: W1213 17:16:55.859550 4989 reflector.go:484] object-"openshift-image-registry"/"node-ca-dockercfg-4777p": watch of *v1.Secret ended with: very short watch: object-"openshift-image-registry"/"node-ca-dockercfg-4777p": Unexpected watch close - watch lasted less than a second and no items received Dec 13 17:16:55 crc kubenswrapper[4989]: W1213 17:16:55.860584 4989 reflector.go:484] object-"openshift-image-registry"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.939933 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.939973 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.939985 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.940002 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:55 crc kubenswrapper[4989]: I1213 17:16:55.940016 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:55Z","lastTransitionTime":"2025-12-13T17:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.014007 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.014041 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.014131 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:56 crc kubenswrapper[4989]: E1213 17:16:56.014268 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:16:56 crc kubenswrapper[4989]: E1213 17:16:56.014619 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:16:56 crc kubenswrapper[4989]: E1213 17:16:56.014731 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.035044 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.044802 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.044841 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.044851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.044871 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.044907 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:56Z","lastTransitionTime":"2025-12-13T17:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.049333 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.064373 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.080820 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.104769 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.118445 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.138864 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.148424 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.148488 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.148498 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.148515 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.148524 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:56Z","lastTransitionTime":"2025-12-13T17:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.153409 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.166447 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.189539 4989 generic.go:334] "Generic (PLEG): container finished" podID="59ed6e54-65fe-4383-9578-d0c89a69ecec" containerID="ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb" exitCode=0 Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.189609 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" event={"ID":"59ed6e54-65fe-4383-9578-d0c89a69ecec","Type":"ContainerDied","Data":"ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.193876 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.195739 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vmx98" event={"ID":"25dd508e-594b-4f32-af84-61ee8d65f38b","Type":"ContainerStarted","Data":"cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.195769 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vmx98" event={"ID":"25dd508e-594b-4f32-af84-61ee8d65f38b","Type":"ContainerStarted","Data":"c95c37c96164687c39042035e0dc90effc36017181ba851b19d91e8ccec0937e"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.204385 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.243117 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.251945 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.251996 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.252006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.252025 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.252035 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:56Z","lastTransitionTime":"2025-12-13T17:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.282091 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.341075 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.360512 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.360552 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.360564 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.360585 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.360598 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:56Z","lastTransitionTime":"2025-12-13T17:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.365480 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.417977 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.439544 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.472739 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.472770 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.472780 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.472815 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.472826 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:56Z","lastTransitionTime":"2025-12-13T17:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.484280 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.522320 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.561481 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.576157 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.576198 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.576206 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.576219 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.576229 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:56Z","lastTransitionTime":"2025-12-13T17:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.603363 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.646445 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.681380 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.681419 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.681428 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.681442 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.681452 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:56Z","lastTransitionTime":"2025-12-13T17:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.683223 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.724097 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.761108 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.783667 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.783692 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.783699 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.783712 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.783721 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:56Z","lastTransitionTime":"2025-12-13T17:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.804712 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.846964 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.884218 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.885638 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.885675 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.885686 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.885703 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.885716 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:56Z","lastTransitionTime":"2025-12-13T17:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.925679 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.966321 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.972971 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.990863 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.990921 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.990935 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.990954 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:56 crc kubenswrapper[4989]: I1213 17:16:56.990967 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:56Z","lastTransitionTime":"2025-12-13T17:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.020256 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.032551 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.052951 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.093735 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.093779 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.093805 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.093824 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.093836 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.197333 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.197413 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.197432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.197460 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.197477 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.204019 4989 generic.go:334] "Generic (PLEG): container finished" podID="59ed6e54-65fe-4383-9578-d0c89a69ecec" containerID="24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a" exitCode=0 Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.204086 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" event={"ID":"59ed6e54-65fe-4383-9578-d0c89a69ecec","Type":"ContainerDied","Data":"24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.236390 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.253779 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.271977 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.285954 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.300111 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.300147 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.300155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.300169 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.300181 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.305840 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.330425 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.345085 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.384762 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.393327 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.402718 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.402758 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.402769 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.402800 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.402814 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.413940 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.413960 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.413967 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.413976 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.413983 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: E1213 17:16:57.427009 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.432341 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.432384 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.432396 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.432414 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.432429 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.443565 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: E1213 17:16:57.445457 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.448656 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.448708 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.448720 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.448739 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.448751 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: E1213 17:16:57.465151 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.471879 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.471923 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.471936 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.471954 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.471967 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.480674 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: E1213 17:16:57.482956 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.486164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.486198 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.486209 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.486226 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.486237 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: E1213 17:16:57.496684 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: E1213 17:16:57.496845 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.505058 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.505103 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.505112 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.505129 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.505141 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.520294 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.561342 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.604368 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.607139 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.607186 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.607278 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.607299 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.607312 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.641081 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.689862 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:57Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.709454 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.709497 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.709510 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.709527 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.709541 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.811891 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.811938 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.811952 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.811972 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.811984 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.914621 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.914657 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.914665 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.914678 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:57 crc kubenswrapper[4989]: I1213 17:16:57.914687 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:57Z","lastTransitionTime":"2025-12-13T17:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.014263 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.014298 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:58 crc kubenswrapper[4989]: E1213 17:16:58.014395 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.014441 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:58 crc kubenswrapper[4989]: E1213 17:16:58.014627 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:16:58 crc kubenswrapper[4989]: E1213 17:16:58.014822 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.018046 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.018080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.018127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.018147 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.018159 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:58Z","lastTransitionTime":"2025-12-13T17:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.120651 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.120702 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.120718 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.120738 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.120754 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:58Z","lastTransitionTime":"2025-12-13T17:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.211074 4989 generic.go:334] "Generic (PLEG): container finished" podID="59ed6e54-65fe-4383-9578-d0c89a69ecec" containerID="67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410" exitCode=0 Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.211138 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" event={"ID":"59ed6e54-65fe-4383-9578-d0c89a69ecec","Type":"ContainerDied","Data":"67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410"} Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.225179 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.225214 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.225226 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.225241 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.225253 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:58Z","lastTransitionTime":"2025-12-13T17:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.227750 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.244693 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.262902 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.280173 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.292700 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.302586 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.316008 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.327910 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.328924 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.328981 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.328992 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.329006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.329017 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:58Z","lastTransitionTime":"2025-12-13T17:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.338804 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.354081 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.381808 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.395280 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.406837 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.418312 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.430872 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.430912 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.430934 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.430951 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.430963 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:58Z","lastTransitionTime":"2025-12-13T17:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.439390 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.533323 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.533347 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.533355 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.533369 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.533378 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:58Z","lastTransitionTime":"2025-12-13T17:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.635560 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.635595 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.635604 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.635618 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.635628 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:58Z","lastTransitionTime":"2025-12-13T17:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.739093 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.739124 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.739135 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.739149 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.739159 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:58Z","lastTransitionTime":"2025-12-13T17:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.842074 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.842330 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.842346 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.842366 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.842380 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:58Z","lastTransitionTime":"2025-12-13T17:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.944748 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.944847 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.944872 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.944903 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:58 crc kubenswrapper[4989]: I1213 17:16:58.944927 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:58Z","lastTransitionTime":"2025-12-13T17:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.048191 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.048267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.048295 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.048328 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.048354 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:59Z","lastTransitionTime":"2025-12-13T17:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.151559 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.151604 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.151614 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.151628 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.151639 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:59Z","lastTransitionTime":"2025-12-13T17:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.219104 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" event={"ID":"59ed6e54-65fe-4383-9578-d0c89a69ecec","Type":"ContainerStarted","Data":"c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.223625 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.224038 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.249138 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.251173 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.254187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.254269 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.254288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.254313 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.254337 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:59Z","lastTransitionTime":"2025-12-13T17:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.264407 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.275320 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.286967 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.302534 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.313458 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.324704 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.335336 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.347290 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.355950 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.357566 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.357593 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.357603 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.357621 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.357631 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:59Z","lastTransitionTime":"2025-12-13T17:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.367882 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.379020 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.389496 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.402006 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.409527 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.425488 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.434518 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.455684 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.464036 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.464077 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.464091 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.464109 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.464122 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:59Z","lastTransitionTime":"2025-12-13T17:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.467466 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.484487 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.495064 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.505731 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.516270 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.525732 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.533808 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.545488 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.556770 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.565716 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.565752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.565765 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.565783 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.565812 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:59Z","lastTransitionTime":"2025-12-13T17:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.566762 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.578097 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.586035 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:16:59Z is after 2025-08-24T17:21:41Z" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.667702 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.667742 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.667752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.667768 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.667778 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:59Z","lastTransitionTime":"2025-12-13T17:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.746091 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.746449 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:17:15.746422496 +0000 UTC m=+50.352869634 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.770871 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.771090 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.771156 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.771216 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.771270 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:59Z","lastTransitionTime":"2025-12-13T17:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.847695 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.847837 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.847894 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.847934 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.847997 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.848031 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.848070 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.848080 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.848085 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.848099 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:15.848068494 +0000 UTC m=+50.454515662 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.848104 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.848140 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.848156 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:15.848134606 +0000 UTC m=+50.454581784 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.848159 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.848182 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:15.848170047 +0000 UTC m=+50.454617225 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:59 crc kubenswrapper[4989]: E1213 17:16:59.848219 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:15.848201008 +0000 UTC m=+50.454648186 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.872992 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.873045 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.873057 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.873078 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.873091 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:59Z","lastTransitionTime":"2025-12-13T17:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.975866 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.975940 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.975963 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.975995 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:16:59 crc kubenswrapper[4989]: I1213 17:16:59.976019 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:16:59Z","lastTransitionTime":"2025-12-13T17:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.014624 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.014694 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:00 crc kubenswrapper[4989]: E1213 17:17:00.014843 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.014878 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:00 crc kubenswrapper[4989]: E1213 17:17:00.014918 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:00 crc kubenswrapper[4989]: E1213 17:17:00.015088 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.079656 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.079712 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.079724 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.079743 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.079757 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:00Z","lastTransitionTime":"2025-12-13T17:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.182897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.182965 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.182988 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.183018 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.183039 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:00Z","lastTransitionTime":"2025-12-13T17:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.227775 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.228476 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.286275 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.286338 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.286356 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.286380 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.286398 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:00Z","lastTransitionTime":"2025-12-13T17:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.305363 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.322868 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.342885 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.377133 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.388651 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.389088 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.389552 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.389773 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.390342 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:00Z","lastTransitionTime":"2025-12-13T17:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.415984 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.439138 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.464206 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.477915 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.489974 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.492449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.492479 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.492488 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.492503 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.492515 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:00Z","lastTransitionTime":"2025-12-13T17:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.502410 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.520051 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.536477 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.551698 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.569102 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.582360 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.595221 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.595267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.595279 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.595300 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.595312 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:00Z","lastTransitionTime":"2025-12-13T17:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.603104 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:00Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.697478 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.697516 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.697525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.697539 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.697548 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:00Z","lastTransitionTime":"2025-12-13T17:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.800023 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.800063 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.800073 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.800088 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.800101 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:00Z","lastTransitionTime":"2025-12-13T17:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.902179 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.902213 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.902225 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.902239 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:00 crc kubenswrapper[4989]: I1213 17:17:00.902248 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:00Z","lastTransitionTime":"2025-12-13T17:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.006481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.006524 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.006535 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.006553 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.006564 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:01Z","lastTransitionTime":"2025-12-13T17:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.110004 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.110081 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.110127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.110153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.110170 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:01Z","lastTransitionTime":"2025-12-13T17:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.213190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.213250 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.213262 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.213281 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.213297 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:01Z","lastTransitionTime":"2025-12-13T17:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.231818 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.315808 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.315860 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.315875 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.315896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.315911 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:01Z","lastTransitionTime":"2025-12-13T17:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.418893 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.418930 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.418940 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.418956 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.418967 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:01Z","lastTransitionTime":"2025-12-13T17:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.521870 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.521922 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.521938 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.521964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.521981 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:01Z","lastTransitionTime":"2025-12-13T17:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.627916 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.627998 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.628016 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.628040 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.628060 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:01Z","lastTransitionTime":"2025-12-13T17:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.730970 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.731017 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.731030 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.731049 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.731061 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:01Z","lastTransitionTime":"2025-12-13T17:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.834288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.834332 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.834341 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.834385 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.834397 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:01Z","lastTransitionTime":"2025-12-13T17:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.937318 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.937400 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.937423 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.937453 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:01 crc kubenswrapper[4989]: I1213 17:17:01.937472 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:01Z","lastTransitionTime":"2025-12-13T17:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.014258 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.014394 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.014413 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:02 crc kubenswrapper[4989]: E1213 17:17:02.014552 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:02 crc kubenswrapper[4989]: E1213 17:17:02.014709 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:02 crc kubenswrapper[4989]: E1213 17:17:02.014906 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.040539 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.040593 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.040609 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.040630 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.040645 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:02Z","lastTransitionTime":"2025-12-13T17:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.142639 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.142677 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.142685 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.142727 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.142738 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:02Z","lastTransitionTime":"2025-12-13T17:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.237934 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/0.log" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.241392 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2" exitCode=1 Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.241433 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2"} Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.242607 4989 scope.go:117] "RemoveContainer" containerID="da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.245120 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.245172 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.245193 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.245221 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.245247 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:02Z","lastTransitionTime":"2025-12-13T17:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.270410 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.286294 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.308983 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.323904 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.340339 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.348018 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.348069 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.348078 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.348094 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.348109 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:02Z","lastTransitionTime":"2025-12-13T17:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.355154 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.368835 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.381652 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.395039 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.414170 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.428428 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.438857 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.448915 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.450507 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.450684 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.450715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.450742 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.450760 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:02Z","lastTransitionTime":"2025-12-13T17:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.469397 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:01Z\\\",\\\"message\\\":\\\"\\\\nI1213 17:17:00.987203 6264 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:00.987214 6264 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:00.987233 6264 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 17:17:00.987247 6264 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1213 17:17:00.987252 6264 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1213 17:17:00.987284 6264 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1213 17:17:00.987938 6264 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:00.987955 6264 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:00.987962 6264 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1213 17:17:00.987973 6264 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 17:17:00.987980 6264 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1213 17:17:00.987980 6264 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 17:17:00.987993 6264 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:00.988004 6264 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:00.988036 6264 factory.go:656] Stopping watch factory\\\\nI1213 17:17:00.988056 6264 ovnkube.go:599] Stopped ovnkube\\\\nI1213 17:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.484653 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:02Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.553403 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.553447 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.553460 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.553481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.553493 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:02Z","lastTransitionTime":"2025-12-13T17:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.655457 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.655501 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.655512 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.655546 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.655558 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:02Z","lastTransitionTime":"2025-12-13T17:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.758410 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.758452 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.758466 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.758489 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.758499 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:02Z","lastTransitionTime":"2025-12-13T17:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.860711 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.860744 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.860851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.860873 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.860885 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:02Z","lastTransitionTime":"2025-12-13T17:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.963608 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.963650 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.963660 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.963674 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:02 crc kubenswrapper[4989]: I1213 17:17:02.963683 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:02Z","lastTransitionTime":"2025-12-13T17:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.066202 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.066244 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.066252 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.066267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.066276 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:03Z","lastTransitionTime":"2025-12-13T17:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.168174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.168220 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.168230 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.168245 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.168425 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:03Z","lastTransitionTime":"2025-12-13T17:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.247769 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/0.log" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.251354 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc"} Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.251442 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.269073 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.270422 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.270464 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.270476 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.270492 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.270504 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:03Z","lastTransitionTime":"2025-12-13T17:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.292269 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:01Z\\\",\\\"message\\\":\\\"\\\\nI1213 17:17:00.987203 6264 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:00.987214 6264 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:00.987233 6264 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 17:17:00.987247 6264 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1213 17:17:00.987252 6264 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1213 17:17:00.987284 6264 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1213 17:17:00.987938 6264 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:00.987955 6264 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:00.987962 6264 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1213 17:17:00.987973 6264 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 17:17:00.987980 6264 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1213 17:17:00.987980 6264 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 17:17:00.987993 6264 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:00.988004 6264 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:00.988036 6264 factory.go:656] Stopping watch factory\\\\nI1213 17:17:00.988056 6264 ovnkube.go:599] Stopped ovnkube\\\\nI1213 17:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.312728 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.326857 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.341165 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.355742 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.376000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.376046 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.376060 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.376078 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.376094 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:03Z","lastTransitionTime":"2025-12-13T17:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.376142 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.387135 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.401039 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.415815 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.429232 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.450754 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.461840 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.475690 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.478005 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.478032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.478040 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.478052 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.478061 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:03Z","lastTransitionTime":"2025-12-13T17:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.490759 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:03Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.581017 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.581090 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.581108 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.581132 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.581150 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:03Z","lastTransitionTime":"2025-12-13T17:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.684717 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.684805 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.684818 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.684836 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.684848 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:03Z","lastTransitionTime":"2025-12-13T17:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.787474 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.787560 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.787676 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.787714 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.787740 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:03Z","lastTransitionTime":"2025-12-13T17:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.890687 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.890759 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.890779 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.890848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.890872 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:03Z","lastTransitionTime":"2025-12-13T17:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.993205 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.993247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.993258 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.993273 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:03 crc kubenswrapper[4989]: I1213 17:17:03.993284 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:03Z","lastTransitionTime":"2025-12-13T17:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.014054 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:04 crc kubenswrapper[4989]: E1213 17:17:04.014165 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.014065 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.014199 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:04 crc kubenswrapper[4989]: E1213 17:17:04.014332 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:04 crc kubenswrapper[4989]: E1213 17:17:04.014578 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.095196 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.095226 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.095234 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.095246 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.095254 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:04Z","lastTransitionTime":"2025-12-13T17:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.198002 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.198034 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.198043 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.198057 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.198070 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:04Z","lastTransitionTime":"2025-12-13T17:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.256352 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/1.log" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.256895 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/0.log" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.259117 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc" exitCode=1 Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.259160 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc"} Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.259197 4989 scope.go:117] "RemoveContainer" containerID="da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.259656 4989 scope.go:117] "RemoveContainer" containerID="5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc" Dec 13 17:17:04 crc kubenswrapper[4989]: E1213 17:17:04.259812 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.277156 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.294335 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.303375 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.303408 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.303418 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.303435 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.303445 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:04Z","lastTransitionTime":"2025-12-13T17:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.311426 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.324562 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.337947 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.350821 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.362933 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.381030 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.393692 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.405223 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.405268 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.405279 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.405298 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.405310 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:04Z","lastTransitionTime":"2025-12-13T17:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.416766 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.432344 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.449949 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.469569 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.501859 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:01Z\\\",\\\"message\\\":\\\"\\\\nI1213 17:17:00.987203 6264 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:00.987214 6264 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:00.987233 6264 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 17:17:00.987247 6264 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1213 17:17:00.987252 6264 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1213 17:17:00.987284 6264 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1213 17:17:00.987938 6264 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:00.987955 6264 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:00.987962 6264 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1213 17:17:00.987973 6264 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 17:17:00.987980 6264 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1213 17:17:00.987980 6264 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 17:17:00.987993 6264 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:00.988004 6264 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:00.988036 6264 factory.go:656] Stopping watch factory\\\\nI1213 17:17:00.988056 6264 ovnkube.go:599] Stopped ovnkube\\\\nI1213 17:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:03Z\\\",\\\"message\\\":\\\"t-go/informers/factory.go:160\\\\nI1213 17:17:03.312552 6417 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.312677 6417 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313034 6417 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:03.313073 6417 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:03.313079 6417 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:03.313127 6417 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 17:17:03.313134 6417 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 17:17:03.313148 6417 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313158 6417 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 17:17:03.313167 6417 factory.go:656] Stopping watch factory\\\\nI1213 17:17:03.313176 6417 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:03.313149 6417 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.508057 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.508085 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.508093 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.508108 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.508117 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:04Z","lastTransitionTime":"2025-12-13T17:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.518376 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.612023 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.612083 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.612100 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.612127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.612148 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:04Z","lastTransitionTime":"2025-12-13T17:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.628124 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h"] Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.629040 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.632192 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.633662 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.652463 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.675546 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.696746 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.711125 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.715546 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.715655 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.715683 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.715720 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.715745 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:04Z","lastTransitionTime":"2025-12-13T17:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.725175 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.740990 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.756661 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.775551 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.792105 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.803984 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.807274 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bf576839-f84e-436f-8855-d0027a0c6ee4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.807320 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwxcs\" (UniqueName: \"kubernetes.io/projected/bf576839-f84e-436f-8855-d0027a0c6ee4-kube-api-access-vwxcs\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.807379 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bf576839-f84e-436f-8855-d0027a0c6ee4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.807421 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bf576839-f84e-436f-8855-d0027a0c6ee4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.818920 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.818954 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.818968 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.818988 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.818999 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:04Z","lastTransitionTime":"2025-12-13T17:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.825121 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.844081 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.861578 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.882389 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.904392 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da4743e9fbb67e2d894845c5d40857dd2eefd54d2619b24352cb0bf09080dcd2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:01Z\\\",\\\"message\\\":\\\"\\\\nI1213 17:17:00.987203 6264 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:00.987214 6264 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:00.987233 6264 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1213 17:17:00.987247 6264 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1213 17:17:00.987252 6264 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1213 17:17:00.987284 6264 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1213 17:17:00.987938 6264 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:00.987955 6264 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:00.987962 6264 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1213 17:17:00.987973 6264 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1213 17:17:00.987980 6264 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1213 17:17:00.987980 6264 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1213 17:17:00.987993 6264 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:00.988004 6264 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:00.988036 6264 factory.go:656] Stopping watch factory\\\\nI1213 17:17:00.988056 6264 ovnkube.go:599] Stopped ovnkube\\\\nI1213 17:17:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:03Z\\\",\\\"message\\\":\\\"t-go/informers/factory.go:160\\\\nI1213 17:17:03.312552 6417 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.312677 6417 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313034 6417 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:03.313073 6417 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:03.313079 6417 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:03.313127 6417 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 17:17:03.313134 6417 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 17:17:03.313148 6417 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313158 6417 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 17:17:03.313167 6417 factory.go:656] Stopping watch factory\\\\nI1213 17:17:03.313176 6417 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:03.313149 6417 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.908147 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bf576839-f84e-436f-8855-d0027a0c6ee4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.908227 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bf576839-f84e-436f-8855-d0027a0c6ee4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.908262 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwxcs\" (UniqueName: \"kubernetes.io/projected/bf576839-f84e-436f-8855-d0027a0c6ee4-kube-api-access-vwxcs\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.908324 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bf576839-f84e-436f-8855-d0027a0c6ee4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.909221 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bf576839-f84e-436f-8855-d0027a0c6ee4-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.909444 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bf576839-f84e-436f-8855-d0027a0c6ee4-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.915697 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bf576839-f84e-436f-8855-d0027a0c6ee4-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.922706 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.922751 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.922759 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.922775 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.922803 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:04Z","lastTransitionTime":"2025-12-13T17:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.922925 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:04Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.930123 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwxcs\" (UniqueName: \"kubernetes.io/projected/bf576839-f84e-436f-8855-d0027a0c6ee4-kube-api-access-vwxcs\") pod \"ovnkube-control-plane-749d76644c-tbb8h\" (UID: \"bf576839-f84e-436f-8855-d0027a0c6ee4\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: I1213 17:17:04.956635 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" Dec 13 17:17:04 crc kubenswrapper[4989]: W1213 17:17:04.970520 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf576839_f84e_436f_8855_d0027a0c6ee4.slice/crio-455673585f72cd8bd855e7944fa7a9e55b26910fec50b37a6a7ed6fe3b8aaa2d WatchSource:0}: Error finding container 455673585f72cd8bd855e7944fa7a9e55b26910fec50b37a6a7ed6fe3b8aaa2d: Status 404 returned error can't find the container with id 455673585f72cd8bd855e7944fa7a9e55b26910fec50b37a6a7ed6fe3b8aaa2d Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.026068 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.026103 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.026111 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.026127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.026136 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:05Z","lastTransitionTime":"2025-12-13T17:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.127947 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.127979 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.127990 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.128006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.128017 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:05Z","lastTransitionTime":"2025-12-13T17:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.231775 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.231815 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.231823 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.231836 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.231844 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:05Z","lastTransitionTime":"2025-12-13T17:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.268171 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" event={"ID":"bf576839-f84e-436f-8855-d0027a0c6ee4","Type":"ContainerStarted","Data":"6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8"} Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.268220 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" event={"ID":"bf576839-f84e-436f-8855-d0027a0c6ee4","Type":"ContainerStarted","Data":"455673585f72cd8bd855e7944fa7a9e55b26910fec50b37a6a7ed6fe3b8aaa2d"} Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.272149 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/1.log" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.276417 4989 scope.go:117] "RemoveContainer" containerID="5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc" Dec 13 17:17:05 crc kubenswrapper[4989]: E1213 17:17:05.276593 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.288661 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.301007 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.323240 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.334514 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.334546 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.334557 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.334573 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.334582 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:05Z","lastTransitionTime":"2025-12-13T17:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.338495 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.356319 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.372005 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.373238 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-lfpf8"] Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.373963 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:05 crc kubenswrapper[4989]: E1213 17:17:05.374054 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.388984 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.412690 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:03Z\\\",\\\"message\\\":\\\"t-go/informers/factory.go:160\\\\nI1213 17:17:03.312552 6417 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.312677 6417 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313034 6417 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:03.313073 6417 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:03.313079 6417 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:03.313127 6417 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 17:17:03.313134 6417 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 17:17:03.313148 6417 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313158 6417 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 17:17:03.313167 6417 factory.go:656] Stopping watch factory\\\\nI1213 17:17:03.313176 6417 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:03.313149 6417 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.437164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.437195 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.437204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.437217 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.437226 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:05Z","lastTransitionTime":"2025-12-13T17:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.444475 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.455234 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.466495 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.483315 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.496020 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.506389 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.514850 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45wzg\" (UniqueName: \"kubernetes.io/projected/7d912915-788e-412a-bae8-6eccd6b4c238-kube-api-access-45wzg\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.514903 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.516896 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.529378 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.539015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.539065 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.539076 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.539088 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.539098 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:05Z","lastTransitionTime":"2025-12-13T17:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.541545 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.554472 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.568926 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.581246 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.597898 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.615994 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.616110 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45wzg\" (UniqueName: \"kubernetes.io/projected/7d912915-788e-412a-bae8-6eccd6b4c238-kube-api-access-45wzg\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:05 crc kubenswrapper[4989]: E1213 17:17:05.616187 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:05 crc kubenswrapper[4989]: E1213 17:17:05.616305 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs podName:7d912915-788e-412a-bae8-6eccd6b4c238 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:06.116270948 +0000 UTC m=+40.722718116 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs") pod "network-metrics-daemon-lfpf8" (UID: "7d912915-788e-412a-bae8-6eccd6b4c238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.620823 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.640764 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.640855 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.640879 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.640903 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.640920 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:05Z","lastTransitionTime":"2025-12-13T17:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.642785 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.645627 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45wzg\" (UniqueName: \"kubernetes.io/projected/7d912915-788e-412a-bae8-6eccd6b4c238-kube-api-access-45wzg\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.657930 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.674882 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.687695 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.710670 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.729528 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.743282 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.743328 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.743343 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.743364 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.743406 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:05Z","lastTransitionTime":"2025-12-13T17:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.743827 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.759625 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.779963 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:03Z\\\",\\\"message\\\":\\\"t-go/informers/factory.go:160\\\\nI1213 17:17:03.312552 6417 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.312677 6417 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313034 6417 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:03.313073 6417 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:03.313079 6417 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:03.313127 6417 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 17:17:03.313134 6417 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 17:17:03.313148 6417 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313158 6417 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 17:17:03.313167 6417 factory.go:656] Stopping watch factory\\\\nI1213 17:17:03.313176 6417 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:03.313149 6417 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.793485 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.807258 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:05Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.849064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.849102 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.849116 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.849140 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.849177 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:05Z","lastTransitionTime":"2025-12-13T17:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.953324 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.953380 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.953397 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.953434 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:05 crc kubenswrapper[4989]: I1213 17:17:05.953451 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:05Z","lastTransitionTime":"2025-12-13T17:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.014439 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.014588 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:06 crc kubenswrapper[4989]: E1213 17:17:06.014695 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.014741 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:06 crc kubenswrapper[4989]: E1213 17:17:06.014862 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:06 crc kubenswrapper[4989]: E1213 17:17:06.015022 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.038922 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.056763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.056883 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.056908 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.056937 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.056960 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:06Z","lastTransitionTime":"2025-12-13T17:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.060458 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.078571 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.102453 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.117853 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.120853 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:06 crc kubenswrapper[4989]: E1213 17:17:06.121016 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:06 crc kubenswrapper[4989]: E1213 17:17:06.121091 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs podName:7d912915-788e-412a-bae8-6eccd6b4c238 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:07.121071329 +0000 UTC m=+41.727518467 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs") pod "network-metrics-daemon-lfpf8" (UID: "7d912915-788e-412a-bae8-6eccd6b4c238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.142740 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.157774 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.159534 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.159584 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.159598 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.159620 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.159638 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:06Z","lastTransitionTime":"2025-12-13T17:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.176940 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.194512 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.212887 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:03Z\\\",\\\"message\\\":\\\"t-go/informers/factory.go:160\\\\nI1213 17:17:03.312552 6417 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.312677 6417 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313034 6417 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:03.313073 6417 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:03.313079 6417 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:03.313127 6417 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 17:17:03.313134 6417 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 17:17:03.313148 6417 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313158 6417 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 17:17:03.313167 6417 factory.go:656] Stopping watch factory\\\\nI1213 17:17:03.313176 6417 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:03.313149 6417 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.228713 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.244925 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.259492 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.262826 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.262870 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.262883 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.262899 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.262911 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:06Z","lastTransitionTime":"2025-12-13T17:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.274706 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.286321 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" event={"ID":"bf576839-f84e-436f-8855-d0027a0c6ee4","Type":"ContainerStarted","Data":"99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b"} Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.289490 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.299699 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.312035 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.326049 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.340335 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.352192 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.366092 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.366089 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.366440 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.366661 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.366685 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.366698 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:06Z","lastTransitionTime":"2025-12-13T17:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.377326 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.413914 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.433554 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.450898 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.461315 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.468563 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.468738 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.468815 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.468880 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.468942 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:06Z","lastTransitionTime":"2025-12-13T17:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.477854 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:03Z\\\",\\\"message\\\":\\\"t-go/informers/factory.go:160\\\\nI1213 17:17:03.312552 6417 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.312677 6417 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313034 6417 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:03.313073 6417 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:03.313079 6417 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:03.313127 6417 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 17:17:03.313134 6417 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 17:17:03.313148 6417 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313158 6417 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 17:17:03.313167 6417 factory.go:656] Stopping watch factory\\\\nI1213 17:17:03.313176 6417 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:03.313149 6417 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.488993 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.502408 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.517021 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.529405 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.543558 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.554373 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.565821 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.571715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.571740 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.571749 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.571763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.571772 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:06Z","lastTransitionTime":"2025-12-13T17:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.673903 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.674127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.674188 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.674257 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.674352 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:06Z","lastTransitionTime":"2025-12-13T17:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.776360 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.776403 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.776412 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.776428 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.776439 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:06Z","lastTransitionTime":"2025-12-13T17:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.879184 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.879379 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.879390 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.879406 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.879416 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:06Z","lastTransitionTime":"2025-12-13T17:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.981887 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.982042 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.982121 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.982189 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:06 crc kubenswrapper[4989]: I1213 17:17:06.982246 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:06Z","lastTransitionTime":"2025-12-13T17:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.013948 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:07 crc kubenswrapper[4989]: E1213 17:17:07.014159 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.085772 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.085890 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.085912 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.085947 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.085967 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.130809 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:07 crc kubenswrapper[4989]: E1213 17:17:07.130995 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:07 crc kubenswrapper[4989]: E1213 17:17:07.131058 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs podName:7d912915-788e-412a-bae8-6eccd6b4c238 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:09.131039073 +0000 UTC m=+43.737486221 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs") pod "network-metrics-daemon-lfpf8" (UID: "7d912915-788e-412a-bae8-6eccd6b4c238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.189906 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.189969 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.189992 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.190019 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.190038 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.293274 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.293332 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.293351 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.293378 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.293397 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.396748 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.396815 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.396827 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.396847 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.396860 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.499530 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.500736 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.501038 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.501209 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.501375 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.603612 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.603673 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.603692 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.603720 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.603740 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.665913 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.665976 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.665990 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.666013 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.666030 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: E1213 17:17:07.681259 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:07Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.692074 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.692447 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.692515 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.692579 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.692634 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: E1213 17:17:07.714486 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:07Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.720402 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.720462 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.720480 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.720505 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.720524 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: E1213 17:17:07.737832 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:07Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.743033 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.743086 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.743098 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.743117 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.743130 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: E1213 17:17:07.760487 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:07Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.764066 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.764097 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.764108 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.764124 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.764137 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: E1213 17:17:07.782628 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:07Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:07 crc kubenswrapper[4989]: E1213 17:17:07.782769 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.785273 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.785312 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.785327 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.785347 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.785360 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.888137 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.888183 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.888197 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.888216 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.888229 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.990322 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.990370 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.990384 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.990403 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:07 crc kubenswrapper[4989]: I1213 17:17:07.990417 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:07Z","lastTransitionTime":"2025-12-13T17:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.013927 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.013934 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:08 crc kubenswrapper[4989]: E1213 17:17:08.014064 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.014205 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:08 crc kubenswrapper[4989]: E1213 17:17:08.014306 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:08 crc kubenswrapper[4989]: E1213 17:17:08.014410 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.093657 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.093695 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.093705 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.093720 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.093731 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:08Z","lastTransitionTime":"2025-12-13T17:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.196453 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.196523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.196546 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.196577 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.196603 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:08Z","lastTransitionTime":"2025-12-13T17:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.298867 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.298914 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.298930 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.298959 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.298971 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:08Z","lastTransitionTime":"2025-12-13T17:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.401580 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.401629 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.401646 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.401666 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.401681 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:08Z","lastTransitionTime":"2025-12-13T17:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.504567 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.504615 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.504632 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.504651 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.504663 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:08Z","lastTransitionTime":"2025-12-13T17:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.607137 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.607224 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.607239 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.607259 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.607277 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:08Z","lastTransitionTime":"2025-12-13T17:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.709953 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.709986 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.709995 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.710012 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.710023 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:08Z","lastTransitionTime":"2025-12-13T17:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.813227 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.813670 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.813909 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.814110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.814293 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:08Z","lastTransitionTime":"2025-12-13T17:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.917459 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.917488 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.917497 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.917511 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:08 crc kubenswrapper[4989]: I1213 17:17:08.917519 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:08Z","lastTransitionTime":"2025-12-13T17:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.013869 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:09 crc kubenswrapper[4989]: E1213 17:17:09.014067 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.020272 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.020336 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.020363 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.020388 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.020410 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:09Z","lastTransitionTime":"2025-12-13T17:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.123170 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.123249 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.123277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.123307 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.123326 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:09Z","lastTransitionTime":"2025-12-13T17:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.148890 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:09 crc kubenswrapper[4989]: E1213 17:17:09.149161 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:09 crc kubenswrapper[4989]: E1213 17:17:09.149258 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs podName:7d912915-788e-412a-bae8-6eccd6b4c238 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:13.14923186 +0000 UTC m=+47.755679038 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs") pod "network-metrics-daemon-lfpf8" (UID: "7d912915-788e-412a-bae8-6eccd6b4c238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.226647 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.226719 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.226746 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.226776 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.226826 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:09Z","lastTransitionTime":"2025-12-13T17:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.330504 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.330912 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.331135 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.331331 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.331499 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:09Z","lastTransitionTime":"2025-12-13T17:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.434318 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.434375 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.434391 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.434414 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.434433 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:09Z","lastTransitionTime":"2025-12-13T17:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.537965 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.538030 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.538045 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.538065 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.538081 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:09Z","lastTransitionTime":"2025-12-13T17:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.640394 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.640439 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.640451 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.640471 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.640483 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:09Z","lastTransitionTime":"2025-12-13T17:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.743320 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.743395 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.743420 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.743450 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.743476 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:09Z","lastTransitionTime":"2025-12-13T17:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.845976 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.846024 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.846036 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.846060 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.846076 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:09Z","lastTransitionTime":"2025-12-13T17:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.949821 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.949859 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.949869 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.949888 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:09 crc kubenswrapper[4989]: I1213 17:17:09.949900 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:09Z","lastTransitionTime":"2025-12-13T17:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.013587 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.013636 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.013719 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:10 crc kubenswrapper[4989]: E1213 17:17:10.013774 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:10 crc kubenswrapper[4989]: E1213 17:17:10.013899 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:10 crc kubenswrapper[4989]: E1213 17:17:10.014040 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.053058 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.053122 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.053142 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.053168 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.053186 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:10Z","lastTransitionTime":"2025-12-13T17:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.155296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.155337 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.155347 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.155361 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.155370 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:10Z","lastTransitionTime":"2025-12-13T17:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.257776 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.257903 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.257921 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.257952 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.257970 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:10Z","lastTransitionTime":"2025-12-13T17:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.360516 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.360566 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.360580 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.360603 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.360618 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:10Z","lastTransitionTime":"2025-12-13T17:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.463832 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.463891 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.463909 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.463934 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.463950 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:10Z","lastTransitionTime":"2025-12-13T17:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.566140 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.566205 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.566223 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.566248 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.566266 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:10Z","lastTransitionTime":"2025-12-13T17:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.669100 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.669153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.669169 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.669192 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.669209 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:10Z","lastTransitionTime":"2025-12-13T17:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.772525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.772597 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.772616 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.772642 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.772668 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:10Z","lastTransitionTime":"2025-12-13T17:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.876389 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.876455 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.876474 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.876499 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.876516 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:10Z","lastTransitionTime":"2025-12-13T17:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.979916 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.979974 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.979998 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.980026 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:10 crc kubenswrapper[4989]: I1213 17:17:10.980052 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:10Z","lastTransitionTime":"2025-12-13T17:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.013846 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:11 crc kubenswrapper[4989]: E1213 17:17:11.014025 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.083289 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.083386 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.083404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.083431 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.083452 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:11Z","lastTransitionTime":"2025-12-13T17:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.187601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.187653 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.187665 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.187685 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.187697 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:11Z","lastTransitionTime":"2025-12-13T17:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.289898 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.289927 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.289938 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.289951 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.289961 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:11Z","lastTransitionTime":"2025-12-13T17:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.392383 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.392413 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.392421 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.392436 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.392446 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:11Z","lastTransitionTime":"2025-12-13T17:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.495506 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.495559 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.495567 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.495581 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.495591 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:11Z","lastTransitionTime":"2025-12-13T17:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.598122 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.598175 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.598195 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.598218 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.598235 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:11Z","lastTransitionTime":"2025-12-13T17:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.701434 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.701483 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.701518 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.701554 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.701576 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:11Z","lastTransitionTime":"2025-12-13T17:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.804617 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.804716 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.804733 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.804755 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.804768 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:11Z","lastTransitionTime":"2025-12-13T17:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.907575 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.907634 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.907651 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.907677 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:11 crc kubenswrapper[4989]: I1213 17:17:11.907697 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:11Z","lastTransitionTime":"2025-12-13T17:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.011175 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.011258 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.011280 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.011310 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.011336 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:12Z","lastTransitionTime":"2025-12-13T17:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.014076 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.014155 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:12 crc kubenswrapper[4989]: E1213 17:17:12.014210 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:12 crc kubenswrapper[4989]: E1213 17:17:12.014310 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.014392 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:12 crc kubenswrapper[4989]: E1213 17:17:12.014659 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.114337 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.114367 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.114376 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.114389 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.114399 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:12Z","lastTransitionTime":"2025-12-13T17:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.217906 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.217952 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.217980 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.217997 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.218007 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:12Z","lastTransitionTime":"2025-12-13T17:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.320997 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.321059 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.321076 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.321097 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.321112 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:12Z","lastTransitionTime":"2025-12-13T17:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.424299 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.424346 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.424360 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.424375 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.424385 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:12Z","lastTransitionTime":"2025-12-13T17:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.527908 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.527950 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.527959 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.527973 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.527982 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:12Z","lastTransitionTime":"2025-12-13T17:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.632417 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.632485 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.632510 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.632540 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.632567 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:12Z","lastTransitionTime":"2025-12-13T17:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.735508 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.735559 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.735576 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.735596 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.735614 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:12Z","lastTransitionTime":"2025-12-13T17:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.842490 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.842530 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.842540 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.842555 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.842566 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:12Z","lastTransitionTime":"2025-12-13T17:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.945316 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.945381 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.945397 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.945417 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:12 crc kubenswrapper[4989]: I1213 17:17:12.945431 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:12Z","lastTransitionTime":"2025-12-13T17:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.013786 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:13 crc kubenswrapper[4989]: E1213 17:17:13.014032 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.047964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.048042 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.048059 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.048083 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.048156 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:13Z","lastTransitionTime":"2025-12-13T17:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.150505 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.150599 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.150624 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.150653 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.150678 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:13Z","lastTransitionTime":"2025-12-13T17:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.191688 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:13 crc kubenswrapper[4989]: E1213 17:17:13.191968 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:13 crc kubenswrapper[4989]: E1213 17:17:13.192091 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs podName:7d912915-788e-412a-bae8-6eccd6b4c238 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:21.19206538 +0000 UTC m=+55.798512548 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs") pod "network-metrics-daemon-lfpf8" (UID: "7d912915-788e-412a-bae8-6eccd6b4c238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.253958 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.254037 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.254069 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.254099 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.254122 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:13Z","lastTransitionTime":"2025-12-13T17:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.356473 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.356531 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.356552 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.356580 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.356602 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:13Z","lastTransitionTime":"2025-12-13T17:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.459615 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.459661 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.459672 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.459691 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.459705 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:13Z","lastTransitionTime":"2025-12-13T17:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.561786 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.561858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.561870 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.561886 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.561897 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:13Z","lastTransitionTime":"2025-12-13T17:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.664884 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.664981 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.664999 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.665023 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.665040 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:13Z","lastTransitionTime":"2025-12-13T17:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.767419 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.767477 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.767486 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.767501 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.767512 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:13Z","lastTransitionTime":"2025-12-13T17:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.870090 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.870143 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.870161 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.870185 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.870201 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:13Z","lastTransitionTime":"2025-12-13T17:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.973758 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.973872 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.973901 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.973928 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:13 crc kubenswrapper[4989]: I1213 17:17:13.973949 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:13Z","lastTransitionTime":"2025-12-13T17:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.014089 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.014129 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:14 crc kubenswrapper[4989]: E1213 17:17:14.014299 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.014462 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:14 crc kubenswrapper[4989]: E1213 17:17:14.014508 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:14 crc kubenswrapper[4989]: E1213 17:17:14.014725 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.077175 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.077223 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.077240 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.077262 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.077279 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:14Z","lastTransitionTime":"2025-12-13T17:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.180114 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.180178 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.180199 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.180223 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.180241 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:14Z","lastTransitionTime":"2025-12-13T17:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.287774 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.287943 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.287975 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.288002 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.288022 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:14Z","lastTransitionTime":"2025-12-13T17:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.391654 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.391756 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.391860 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.391889 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.391945 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:14Z","lastTransitionTime":"2025-12-13T17:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.494833 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.494896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.494914 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.494938 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.494955 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:14Z","lastTransitionTime":"2025-12-13T17:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.597441 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.597483 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.597492 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.597508 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.597518 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:14Z","lastTransitionTime":"2025-12-13T17:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.699983 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.700046 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.700064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.700089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.700108 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:14Z","lastTransitionTime":"2025-12-13T17:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.802817 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.802852 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.802861 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.802875 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.802883 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:14Z","lastTransitionTime":"2025-12-13T17:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.906187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.906300 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.906323 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.906384 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:14 crc kubenswrapper[4989]: I1213 17:17:14.906403 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:14Z","lastTransitionTime":"2025-12-13T17:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.009525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.009571 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.009588 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.009612 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.009629 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:15Z","lastTransitionTime":"2025-12-13T17:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.014133 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.014424 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.112837 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.112900 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.112923 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.112951 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.112971 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:15Z","lastTransitionTime":"2025-12-13T17:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.215912 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.215958 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.215970 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.215986 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.215998 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:15Z","lastTransitionTime":"2025-12-13T17:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.318314 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.318522 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.318548 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.318577 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.318596 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:15Z","lastTransitionTime":"2025-12-13T17:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.421305 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.421367 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.421387 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.421410 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.421426 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:15Z","lastTransitionTime":"2025-12-13T17:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.524350 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.524394 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.524403 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.524417 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.524427 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:15Z","lastTransitionTime":"2025-12-13T17:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.631947 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.631997 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.632009 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.632028 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.632040 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:15Z","lastTransitionTime":"2025-12-13T17:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.734452 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.734493 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.734506 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.734523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.734535 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:15Z","lastTransitionTime":"2025-12-13T17:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.837027 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.837065 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.837076 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.837091 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.837102 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:15Z","lastTransitionTime":"2025-12-13T17:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.838537 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.838691 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:17:47.838672654 +0000 UTC m=+82.445119792 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.939449 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.939635 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.939683 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.939714 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.939726 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.939839 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.939852 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.939884 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.939912 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.939914 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.939785 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.939856 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.939915 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:47.93989211 +0000 UTC m=+82.546339258 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.940119 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:47.940060345 +0000 UTC m=+82.546507523 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.940172 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:47.940149087 +0000 UTC m=+82.546596325 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:17:15 crc kubenswrapper[4989]: E1213 17:17:15.940202 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:47.940184978 +0000 UTC m=+82.546632196 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.940278 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.940332 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.940349 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.940373 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:15 crc kubenswrapper[4989]: I1213 17:17:15.940390 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:15Z","lastTransitionTime":"2025-12-13T17:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.014329 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:16 crc kubenswrapper[4989]: E1213 17:17:16.014553 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.014582 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:16 crc kubenswrapper[4989]: E1213 17:17:16.014893 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.015012 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:16 crc kubenswrapper[4989]: E1213 17:17:16.015290 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.036048 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.042446 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.042477 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.042488 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.042506 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.042517 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:16Z","lastTransitionTime":"2025-12-13T17:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.050421 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.066492 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.086999 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.101984 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.125115 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.139078 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.144615 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.144654 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.144667 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.144683 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.144697 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:16Z","lastTransitionTime":"2025-12-13T17:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.151519 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.163557 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.179253 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:03Z\\\",\\\"message\\\":\\\"t-go/informers/factory.go:160\\\\nI1213 17:17:03.312552 6417 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.312677 6417 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313034 6417 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:03.313073 6417 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:03.313079 6417 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:03.313127 6417 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 17:17:03.313134 6417 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 17:17:03.313148 6417 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313158 6417 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 17:17:03.313167 6417 factory.go:656] Stopping watch factory\\\\nI1213 17:17:03.313176 6417 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:03.313149 6417 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.190735 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.202578 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.216755 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.228731 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.241274 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.246848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.246879 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.246887 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.246899 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.246910 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:16Z","lastTransitionTime":"2025-12-13T17:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.251463 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.261132 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:16Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.348897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.348964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.348976 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.348991 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.349004 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:16Z","lastTransitionTime":"2025-12-13T17:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.451158 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.451196 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.451213 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.451234 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.451249 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:16Z","lastTransitionTime":"2025-12-13T17:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.552977 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.553001 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.553008 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.553022 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.553031 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:16Z","lastTransitionTime":"2025-12-13T17:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.655340 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.655384 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.655394 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.655409 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.655419 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:16Z","lastTransitionTime":"2025-12-13T17:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.758651 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.758752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.758770 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.758824 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.758847 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:16Z","lastTransitionTime":"2025-12-13T17:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.861982 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.862572 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.862671 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.862705 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.862735 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:16Z","lastTransitionTime":"2025-12-13T17:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.966013 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.966064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.966075 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.966090 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:16 crc kubenswrapper[4989]: I1213 17:17:16.966100 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:16Z","lastTransitionTime":"2025-12-13T17:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.013714 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:17 crc kubenswrapper[4989]: E1213 17:17:17.013902 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.068622 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.068670 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.068681 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.068697 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.068708 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:17Z","lastTransitionTime":"2025-12-13T17:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.171190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.171250 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.171267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.171291 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.171308 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:17Z","lastTransitionTime":"2025-12-13T17:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.274081 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.274155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.274177 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.274205 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.274223 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:17Z","lastTransitionTime":"2025-12-13T17:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.377036 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.377420 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.377594 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.377749 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.377927 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:17Z","lastTransitionTime":"2025-12-13T17:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.481092 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.481163 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.481187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.481215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.481237 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:17Z","lastTransitionTime":"2025-12-13T17:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.585182 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.585232 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.585242 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.585258 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.585269 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:17Z","lastTransitionTime":"2025-12-13T17:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.687087 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.687136 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.687196 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.687215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.687229 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:17Z","lastTransitionTime":"2025-12-13T17:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.790215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.790281 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.790300 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.790328 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.790346 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:17Z","lastTransitionTime":"2025-12-13T17:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.893228 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.893274 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.893285 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.893302 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.893315 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:17Z","lastTransitionTime":"2025-12-13T17:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.973687 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.973753 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.973771 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.973833 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.973860 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:17Z","lastTransitionTime":"2025-12-13T17:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:17 crc kubenswrapper[4989]: E1213 17:17:17.991511 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:17Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.996096 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.996207 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.996235 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.996264 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:17 crc kubenswrapper[4989]: I1213 17:17:17.996291 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:17Z","lastTransitionTime":"2025-12-13T17:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: E1213 17:17:18.012509 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:18Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.013878 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.013917 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.013893 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:18 crc kubenswrapper[4989]: E1213 17:17:18.014319 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:18 crc kubenswrapper[4989]: E1213 17:17:18.014119 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:18 crc kubenswrapper[4989]: E1213 17:17:18.014616 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.017268 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.017305 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.017320 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.017339 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.017357 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: E1213 17:17:18.031931 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:18Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.036085 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.036155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.036182 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.036213 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.036242 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: E1213 17:17:18.049973 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:18Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.054214 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.054238 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.054246 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.054271 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.054284 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: E1213 17:17:18.069500 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:18Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:18 crc kubenswrapper[4989]: E1213 17:17:18.069633 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.071805 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.072010 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.072019 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.072029 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.072040 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.175010 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.175051 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.175061 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.175077 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.175089 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.278048 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.278083 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.278092 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.278104 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.278112 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.381064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.381114 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.381124 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.381142 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.381153 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.392128 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.393101 4989 scope.go:117] "RemoveContainer" containerID="5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.484105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.484155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.484167 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.484187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.484198 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.586905 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.586941 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.586951 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.586966 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.586976 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.692954 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.692995 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.693006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.693022 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.693033 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.796887 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.796930 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.796944 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.796964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.796978 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.900429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.900476 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.900486 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.900504 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:18 crc kubenswrapper[4989]: I1213 17:17:18.900517 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:18Z","lastTransitionTime":"2025-12-13T17:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.002278 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.002318 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.002327 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.002344 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.002355 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:19Z","lastTransitionTime":"2025-12-13T17:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.013435 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:19 crc kubenswrapper[4989]: E1213 17:17:19.013544 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.104435 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.104470 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.104477 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.104491 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.104501 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:19Z","lastTransitionTime":"2025-12-13T17:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.206550 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.206592 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.206609 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.206630 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.206642 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:19Z","lastTransitionTime":"2025-12-13T17:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.309047 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.309100 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.309118 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.309141 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.309158 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:19Z","lastTransitionTime":"2025-12-13T17:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.345025 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/2.log" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.346128 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/1.log" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.348741 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea" exitCode=1 Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.348777 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea"} Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.348823 4989 scope.go:117] "RemoveContainer" containerID="5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.349539 4989 scope.go:117] "RemoveContainer" containerID="836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea" Dec 13 17:17:19 crc kubenswrapper[4989]: E1213 17:17:19.349674 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.368171 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.381142 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.391784 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.401186 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.411230 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.411258 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.411268 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.411285 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.411295 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:19Z","lastTransitionTime":"2025-12-13T17:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.411732 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.423285 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.435703 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.448087 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.469216 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.481145 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.497536 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.509246 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.513089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.513126 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.513138 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.513156 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.513168 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:19Z","lastTransitionTime":"2025-12-13T17:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.522280 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.545641 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:03Z\\\",\\\"message\\\":\\\"t-go/informers/factory.go:160\\\\nI1213 17:17:03.312552 6417 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.312677 6417 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313034 6417 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:03.313073 6417 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:03.313079 6417 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:03.313127 6417 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 17:17:03.313134 6417 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 17:17:03.313148 6417 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313158 6417 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 17:17:03.313167 6417 factory.go:656] Stopping watch factory\\\\nI1213 17:17:03.313176 6417 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:03.313149 6417 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:19Z\\\",\\\"message\\\":\\\"r] map[service.beta.openshift.io/serving-cert-secret-name:canary-serving-cert] [{apps/v1 daemonset ingress-canary f5a2759b-dc3c-483d-93f0-055bac962b12 0xc000659f87 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:8443-tcp,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:8888-tcp,Protocol:TCP,Port:8888,TargetPort:{0 8888 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{ingresscanary.operator.openshift.io/daemonset-ingresscanary: canary_controller,},ClusterIP:10.217.5.34,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.34],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1213 17:17:19.137223 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, hand\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.559234 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.576826 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.590559 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:19Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.615674 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.615725 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.615740 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.615759 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.615775 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:19Z","lastTransitionTime":"2025-12-13T17:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.718753 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.718816 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.718828 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.718843 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.718855 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:19Z","lastTransitionTime":"2025-12-13T17:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.821872 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.821923 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.821940 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.821965 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.821984 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:19Z","lastTransitionTime":"2025-12-13T17:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.924943 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.924989 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.925004 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.925023 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:19 crc kubenswrapper[4989]: I1213 17:17:19.925036 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:19Z","lastTransitionTime":"2025-12-13T17:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.013721 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.013782 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:20 crc kubenswrapper[4989]: E1213 17:17:20.013898 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.013721 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:20 crc kubenswrapper[4989]: E1213 17:17:20.014023 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:20 crc kubenswrapper[4989]: E1213 17:17:20.014102 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.027561 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.027650 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.027666 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.027684 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.027696 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:20Z","lastTransitionTime":"2025-12-13T17:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.131386 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.131425 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.131434 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.131450 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.131459 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:20Z","lastTransitionTime":"2025-12-13T17:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.234303 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.234384 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.234410 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.234444 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.234470 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:20Z","lastTransitionTime":"2025-12-13T17:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.338140 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.338187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.338204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.338226 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.338243 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:20Z","lastTransitionTime":"2025-12-13T17:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.354077 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/2.log" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.441141 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.441186 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.441197 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.441214 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.441226 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:20Z","lastTransitionTime":"2025-12-13T17:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.543574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.543601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.543608 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.543620 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.543628 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:20Z","lastTransitionTime":"2025-12-13T17:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.646375 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.646415 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.646425 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.646441 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.646451 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:20Z","lastTransitionTime":"2025-12-13T17:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.689514 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.698576 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.710014 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.721835 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.731731 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.741602 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.748109 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.748159 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.748176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.748200 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.748217 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:20Z","lastTransitionTime":"2025-12-13T17:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.758541 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:03Z\\\",\\\"message\\\":\\\"t-go/informers/factory.go:160\\\\nI1213 17:17:03.312552 6417 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.312677 6417 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313034 6417 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:03.313073 6417 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:03.313079 6417 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:03.313127 6417 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 17:17:03.313134 6417 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 17:17:03.313148 6417 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313158 6417 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 17:17:03.313167 6417 factory.go:656] Stopping watch factory\\\\nI1213 17:17:03.313176 6417 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:03.313149 6417 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:19Z\\\",\\\"message\\\":\\\"r] map[service.beta.openshift.io/serving-cert-secret-name:canary-serving-cert] [{apps/v1 daemonset ingress-canary f5a2759b-dc3c-483d-93f0-055bac962b12 0xc000659f87 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:8443-tcp,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:8888-tcp,Protocol:TCP,Port:8888,TargetPort:{0 8888 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{ingresscanary.operator.openshift.io/daemonset-ingresscanary: canary_controller,},ClusterIP:10.217.5.34,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.34],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1213 17:17:19.137223 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, hand\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.770229 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.785844 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.802695 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.814820 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.825871 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.836462 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.846679 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.852760 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.852818 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.852831 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.852849 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.852862 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:20Z","lastTransitionTime":"2025-12-13T17:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.859841 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.873366 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.885375 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.899167 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.908477 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:20Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.955960 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.956007 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.956020 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.956038 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:20 crc kubenswrapper[4989]: I1213 17:17:20.956050 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:20Z","lastTransitionTime":"2025-12-13T17:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.014271 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:21 crc kubenswrapper[4989]: E1213 17:17:21.014402 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.058500 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.058541 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.058549 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.058562 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.058573 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:21Z","lastTransitionTime":"2025-12-13T17:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.160410 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.160452 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.160463 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.160481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.160493 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:21Z","lastTransitionTime":"2025-12-13T17:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.193685 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:21 crc kubenswrapper[4989]: E1213 17:17:21.193930 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:21 crc kubenswrapper[4989]: E1213 17:17:21.194020 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs podName:7d912915-788e-412a-bae8-6eccd6b4c238 nodeName:}" failed. No retries permitted until 2025-12-13 17:17:37.193998523 +0000 UTC m=+71.800445741 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs") pod "network-metrics-daemon-lfpf8" (UID: "7d912915-788e-412a-bae8-6eccd6b4c238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.263336 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.263389 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.263407 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.263428 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.263441 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:21Z","lastTransitionTime":"2025-12-13T17:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.365981 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.366019 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.366029 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.366043 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.366054 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:21Z","lastTransitionTime":"2025-12-13T17:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.468135 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.468166 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.468174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.468187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.468195 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:21Z","lastTransitionTime":"2025-12-13T17:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.572525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.572591 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.572615 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.572644 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.572666 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:21Z","lastTransitionTime":"2025-12-13T17:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.675884 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.675946 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.675971 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.676001 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.676069 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:21Z","lastTransitionTime":"2025-12-13T17:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.778508 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.778590 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.778684 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.778715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.778737 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:21Z","lastTransitionTime":"2025-12-13T17:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.881164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.881224 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.881247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.881275 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.881298 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:21Z","lastTransitionTime":"2025-12-13T17:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.983190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.983267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.983279 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.983292 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:21 crc kubenswrapper[4989]: I1213 17:17:21.983301 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:21Z","lastTransitionTime":"2025-12-13T17:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.014458 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:22 crc kubenswrapper[4989]: E1213 17:17:22.014567 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.014466 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.014616 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:22 crc kubenswrapper[4989]: E1213 17:17:22.014680 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:22 crc kubenswrapper[4989]: E1213 17:17:22.014860 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.085732 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.085780 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.085822 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.085840 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.085852 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:22Z","lastTransitionTime":"2025-12-13T17:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.188079 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.188143 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.188161 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.188187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.188207 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:22Z","lastTransitionTime":"2025-12-13T17:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.290231 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.290269 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.290278 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.290295 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.290306 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:22Z","lastTransitionTime":"2025-12-13T17:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.393994 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.394046 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.394065 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.394090 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.394108 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:22Z","lastTransitionTime":"2025-12-13T17:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.496946 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.497319 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.497560 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.497764 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.497994 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:22Z","lastTransitionTime":"2025-12-13T17:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.600935 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.601175 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.601245 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.601303 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.601360 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:22Z","lastTransitionTime":"2025-12-13T17:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.704487 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.704538 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.704554 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.704578 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.704596 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:22Z","lastTransitionTime":"2025-12-13T17:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.808287 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.808652 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.808783 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.809001 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.809136 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:22Z","lastTransitionTime":"2025-12-13T17:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.912125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.912210 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.912228 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.912252 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:22 crc kubenswrapper[4989]: I1213 17:17:22.912268 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:22Z","lastTransitionTime":"2025-12-13T17:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.013696 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:23 crc kubenswrapper[4989]: E1213 17:17:23.014225 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.015242 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.015292 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.015315 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.015341 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.015362 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:23Z","lastTransitionTime":"2025-12-13T17:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.119017 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.119081 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.119090 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.119102 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.119111 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:23Z","lastTransitionTime":"2025-12-13T17:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.222043 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.222096 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.222113 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.222139 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.222202 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:23Z","lastTransitionTime":"2025-12-13T17:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.324777 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.324829 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.324838 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.324854 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.324864 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:23Z","lastTransitionTime":"2025-12-13T17:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.427595 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.427981 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.428005 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.428032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.428054 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:23Z","lastTransitionTime":"2025-12-13T17:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.531120 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.531187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.531241 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.531257 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.531266 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:23Z","lastTransitionTime":"2025-12-13T17:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.633539 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.633589 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.633600 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.633618 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.633630 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:23Z","lastTransitionTime":"2025-12-13T17:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.736217 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.736245 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.736254 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.736266 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.736278 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:23Z","lastTransitionTime":"2025-12-13T17:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.839409 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.839459 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.839477 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.839517 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.839536 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:23Z","lastTransitionTime":"2025-12-13T17:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.941683 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.941721 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.941729 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.941743 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:23 crc kubenswrapper[4989]: I1213 17:17:23.941754 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:23Z","lastTransitionTime":"2025-12-13T17:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.014099 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.014121 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.014293 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:24 crc kubenswrapper[4989]: E1213 17:17:24.014288 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:24 crc kubenswrapper[4989]: E1213 17:17:24.014387 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:24 crc kubenswrapper[4989]: E1213 17:17:24.014461 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.043671 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.043714 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.043723 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.043736 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.043745 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:24Z","lastTransitionTime":"2025-12-13T17:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.145677 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.145710 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.145728 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.145746 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.145757 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:24Z","lastTransitionTime":"2025-12-13T17:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.247404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.247566 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.247583 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.247604 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.247614 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:24Z","lastTransitionTime":"2025-12-13T17:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.349962 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.350027 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.350045 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.350071 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.350088 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:24Z","lastTransitionTime":"2025-12-13T17:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.452780 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.452850 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.452866 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.452887 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.452902 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:24Z","lastTransitionTime":"2025-12-13T17:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.555761 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.555834 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.555847 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.555865 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.555878 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:24Z","lastTransitionTime":"2025-12-13T17:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.657950 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.657991 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.658051 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.658074 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.658090 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:24Z","lastTransitionTime":"2025-12-13T17:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.761046 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.761094 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.761105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.761121 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.761136 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:24Z","lastTransitionTime":"2025-12-13T17:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.863687 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.863732 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.863748 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.863767 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.863780 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:24Z","lastTransitionTime":"2025-12-13T17:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.966043 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.966084 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.966093 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.966111 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:24 crc kubenswrapper[4989]: I1213 17:17:24.966120 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:24Z","lastTransitionTime":"2025-12-13T17:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.014360 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:25 crc kubenswrapper[4989]: E1213 17:17:25.014488 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.069269 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.069327 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.069351 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.069380 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.069403 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:25Z","lastTransitionTime":"2025-12-13T17:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.172701 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.172760 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.172775 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.172836 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.172861 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:25Z","lastTransitionTime":"2025-12-13T17:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.275436 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.275482 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.275523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.275544 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.275558 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:25Z","lastTransitionTime":"2025-12-13T17:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.378183 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.378224 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.378235 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.378255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.378267 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:25Z","lastTransitionTime":"2025-12-13T17:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.481263 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.481329 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.481347 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.481373 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.481395 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:25Z","lastTransitionTime":"2025-12-13T17:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.584745 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.584849 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.584870 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.584899 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.584920 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:25Z","lastTransitionTime":"2025-12-13T17:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.687752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.687898 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.687926 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.687960 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.687985 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:25Z","lastTransitionTime":"2025-12-13T17:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.790778 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.790887 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.790908 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.790933 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.790950 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:25Z","lastTransitionTime":"2025-12-13T17:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.893723 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.893830 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.893860 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.893888 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.893911 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:25Z","lastTransitionTime":"2025-12-13T17:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.995990 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.996053 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.996110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.996137 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:25 crc kubenswrapper[4989]: I1213 17:17:25.996154 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:25Z","lastTransitionTime":"2025-12-13T17:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.015065 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.015153 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.015063 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:26 crc kubenswrapper[4989]: E1213 17:17:26.015254 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:26 crc kubenswrapper[4989]: E1213 17:17:26.015411 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:26 crc kubenswrapper[4989]: E1213 17:17:26.016079 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.041188 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.067785 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.087350 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.099395 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.099451 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.099475 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.099506 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.099528 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:26Z","lastTransitionTime":"2025-12-13T17:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.109519 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.133517 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.149273 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.170146 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ac58345785a19e6edac91b2aa0f67d648e690aed7fba6eb506a899951dbbddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:03Z\\\",\\\"message\\\":\\\"t-go/informers/factory.go:160\\\\nI1213 17:17:03.312552 6417 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.312677 6417 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313034 6417 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1213 17:17:03.313073 6417 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1213 17:17:03.313079 6417 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1213 17:17:03.313127 6417 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1213 17:17:03.313134 6417 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1213 17:17:03.313148 6417 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1213 17:17:03.313158 6417 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1213 17:17:03.313167 6417 factory.go:656] Stopping watch factory\\\\nI1213 17:17:03.313176 6417 handler.go:208] Removed *v1.Node event handler 2\\\\nI1213 17:17:03.313149 6417 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Node event handler 7\\\\nI1213 17:17:03.313194 6417 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:19Z\\\",\\\"message\\\":\\\"r] map[service.beta.openshift.io/serving-cert-secret-name:canary-serving-cert] [{apps/v1 daemonset ingress-canary f5a2759b-dc3c-483d-93f0-055bac962b12 0xc000659f87 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:8443-tcp,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:8888-tcp,Protocol:TCP,Port:8888,TargetPort:{0 8888 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{ingresscanary.operator.openshift.io/daemonset-ingresscanary: canary_controller,},ClusterIP:10.217.5.34,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.34],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1213 17:17:19.137223 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, hand\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.186756 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.203018 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.203064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.203082 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.203105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.203121 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:26Z","lastTransitionTime":"2025-12-13T17:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.220728 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.240730 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.256940 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.274683 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.290529 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.302389 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.305532 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.305571 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.305583 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.305601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.305613 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:26Z","lastTransitionTime":"2025-12-13T17:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.317345 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.332376 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.344474 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30798eda-9ae1-43fd-918b-a13c4be89568\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.360016 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:26Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.410113 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.410148 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.410159 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.410175 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.410187 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:26Z","lastTransitionTime":"2025-12-13T17:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.512873 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.512907 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.512916 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.512929 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.512939 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:26Z","lastTransitionTime":"2025-12-13T17:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.615538 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.615574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.615583 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.615596 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.615605 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:26Z","lastTransitionTime":"2025-12-13T17:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.717690 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.717750 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.717760 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.717772 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.717782 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:26Z","lastTransitionTime":"2025-12-13T17:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.820231 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.820297 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.820319 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.820352 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.820372 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:26Z","lastTransitionTime":"2025-12-13T17:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.923284 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.923660 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.923677 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.923702 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:26 crc kubenswrapper[4989]: I1213 17:17:26.923722 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:26Z","lastTransitionTime":"2025-12-13T17:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.014056 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:27 crc kubenswrapper[4989]: E1213 17:17:27.014195 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.025768 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.025865 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.025889 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.025921 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.025942 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:27Z","lastTransitionTime":"2025-12-13T17:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.128706 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.128770 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.128812 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.128834 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.128846 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:27Z","lastTransitionTime":"2025-12-13T17:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.231525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.231559 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.231569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.231584 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.231593 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:27Z","lastTransitionTime":"2025-12-13T17:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.333627 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.333674 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.333685 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.333704 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.333716 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:27Z","lastTransitionTime":"2025-12-13T17:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.436590 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.436626 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.436634 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.436648 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.436656 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:27Z","lastTransitionTime":"2025-12-13T17:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.539025 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.539055 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.539063 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.539077 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.539085 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:27Z","lastTransitionTime":"2025-12-13T17:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.641536 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.641571 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.641580 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.641592 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.641601 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:27Z","lastTransitionTime":"2025-12-13T17:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.744560 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.744615 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.744629 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.744652 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.744666 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:27Z","lastTransitionTime":"2025-12-13T17:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.847221 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.847307 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.847325 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.847353 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.847371 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:27Z","lastTransitionTime":"2025-12-13T17:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.950266 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.950367 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.950397 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.950438 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:27 crc kubenswrapper[4989]: I1213 17:17:27.950465 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:27Z","lastTransitionTime":"2025-12-13T17:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.014163 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.014235 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:28 crc kubenswrapper[4989]: E1213 17:17:28.014318 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:28 crc kubenswrapper[4989]: E1213 17:17:28.014527 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.014918 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:28 crc kubenswrapper[4989]: E1213 17:17:28.015018 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.053244 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.053292 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.053307 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.053326 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.053341 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.155944 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.155996 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.156011 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.156041 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.156055 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.230690 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.230741 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.230757 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.230778 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.230827 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: E1213 17:17:28.249458 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:28Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.254172 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.254248 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.254276 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.254304 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.254325 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: E1213 17:17:28.273055 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:28Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.277262 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.277306 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.277321 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.277341 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.277357 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: E1213 17:17:28.296982 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:28Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.307223 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.307302 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.307322 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.307355 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.307377 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: E1213 17:17:28.323392 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:28Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.327786 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.327845 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.327858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.327877 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.327890 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: E1213 17:17:28.348231 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:28Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:28 crc kubenswrapper[4989]: E1213 17:17:28.348450 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.350125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.350187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.350207 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.350231 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.350250 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.453627 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.453738 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.453806 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.453846 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.453866 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.557226 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.557289 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.557301 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.557321 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.557334 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.661052 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.661127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.661146 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.661174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.661193 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.764060 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.764116 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.764129 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.764151 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.764165 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.867693 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.867776 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.867841 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.867869 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.867893 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.970495 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.970569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.970584 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.970610 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:28 crc kubenswrapper[4989]: I1213 17:17:28.970624 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:28Z","lastTransitionTime":"2025-12-13T17:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.014505 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:29 crc kubenswrapper[4989]: E1213 17:17:29.014721 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.074062 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.074136 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.074152 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.074174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.074189 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:29Z","lastTransitionTime":"2025-12-13T17:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.176934 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.176973 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.176986 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.177000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.177009 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:29Z","lastTransitionTime":"2025-12-13T17:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.279270 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.279308 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.279318 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.279332 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.279342 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:29Z","lastTransitionTime":"2025-12-13T17:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.381025 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.381086 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.381095 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.381111 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.381121 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:29Z","lastTransitionTime":"2025-12-13T17:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.483584 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.483631 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.483640 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.483656 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.483665 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:29Z","lastTransitionTime":"2025-12-13T17:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.585337 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.585372 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.585380 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.585393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.585401 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:29Z","lastTransitionTime":"2025-12-13T17:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.687321 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.687370 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.687381 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.687401 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.687414 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:29Z","lastTransitionTime":"2025-12-13T17:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.789867 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.789956 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.789971 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.789991 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.790005 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:29Z","lastTransitionTime":"2025-12-13T17:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.892464 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.892511 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.892520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.892536 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.892546 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:29Z","lastTransitionTime":"2025-12-13T17:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.994666 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.994695 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.994703 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.994715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:29 crc kubenswrapper[4989]: I1213 17:17:29.994723 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:29Z","lastTransitionTime":"2025-12-13T17:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.013589 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.013624 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.013715 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:30 crc kubenswrapper[4989]: E1213 17:17:30.013828 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:30 crc kubenswrapper[4989]: E1213 17:17:30.013949 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:30 crc kubenswrapper[4989]: E1213 17:17:30.014313 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.014579 4989 scope.go:117] "RemoveContainer" containerID="836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea" Dec 13 17:17:30 crc kubenswrapper[4989]: E1213 17:17:30.014755 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.027516 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.040941 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.053452 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.064205 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.080517 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.092948 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30798eda-9ae1-43fd-918b-a13c4be89568\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.105953 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.105988 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.106000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.106015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.106026 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:30Z","lastTransitionTime":"2025-12-13T17:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.111382 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.123070 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.136561 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.146807 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.158546 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.171281 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.183209 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.204341 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:19Z\\\",\\\"message\\\":\\\"r] map[service.beta.openshift.io/serving-cert-secret-name:canary-serving-cert] [{apps/v1 daemonset ingress-canary f5a2759b-dc3c-483d-93f0-055bac962b12 0xc000659f87 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:8443-tcp,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:8888-tcp,Protocol:TCP,Port:8888,TargetPort:{0 8888 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{ingresscanary.operator.openshift.io/daemonset-ingresscanary: canary_controller,},ClusterIP:10.217.5.34,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.34],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1213 17:17:19.137223 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, hand\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.207822 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.207865 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.207881 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.207902 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.207917 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:30Z","lastTransitionTime":"2025-12-13T17:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.215321 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.235688 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.248105 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.259321 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:30Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.310420 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.310526 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.310542 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.310565 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.310577 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:30Z","lastTransitionTime":"2025-12-13T17:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.413314 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.413368 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.413378 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.413397 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.413409 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:30Z","lastTransitionTime":"2025-12-13T17:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.515858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.515925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.515942 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.515967 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.515983 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:30Z","lastTransitionTime":"2025-12-13T17:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.618339 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.618401 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.618422 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.618451 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.618473 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:30Z","lastTransitionTime":"2025-12-13T17:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.721151 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.721245 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.721301 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.721328 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.721344 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:30Z","lastTransitionTime":"2025-12-13T17:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.824623 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.824673 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.824682 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.824695 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.824704 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:30Z","lastTransitionTime":"2025-12-13T17:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.928089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.928158 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.928176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.928205 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:30 crc kubenswrapper[4989]: I1213 17:17:30.928223 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:30Z","lastTransitionTime":"2025-12-13T17:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.014197 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:31 crc kubenswrapper[4989]: E1213 17:17:31.014352 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.030445 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.030495 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.030506 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.030523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.030536 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:31Z","lastTransitionTime":"2025-12-13T17:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.133480 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.133510 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.133520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.133535 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.133545 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:31Z","lastTransitionTime":"2025-12-13T17:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.236150 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.236218 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.236237 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.236257 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.236271 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:31Z","lastTransitionTime":"2025-12-13T17:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.338674 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.338731 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.338752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.338777 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.338821 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:31Z","lastTransitionTime":"2025-12-13T17:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.441445 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.441491 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.441503 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.441522 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.441535 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:31Z","lastTransitionTime":"2025-12-13T17:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.543769 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.543856 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.543874 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.543897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.543918 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:31Z","lastTransitionTime":"2025-12-13T17:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.646141 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.646183 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.646192 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.646206 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.646216 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:31Z","lastTransitionTime":"2025-12-13T17:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.749382 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.749418 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.749428 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.749442 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.749451 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:31Z","lastTransitionTime":"2025-12-13T17:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.852729 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.852788 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.852911 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.852934 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.852953 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:31Z","lastTransitionTime":"2025-12-13T17:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.954799 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.954860 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.954869 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.954883 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:31 crc kubenswrapper[4989]: I1213 17:17:31.954894 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:31Z","lastTransitionTime":"2025-12-13T17:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.013925 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.014017 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:32 crc kubenswrapper[4989]: E1213 17:17:32.014086 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.013925 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:32 crc kubenswrapper[4989]: E1213 17:17:32.014137 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:32 crc kubenswrapper[4989]: E1213 17:17:32.014176 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.057643 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.057696 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.057706 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.057727 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.057739 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:32Z","lastTransitionTime":"2025-12-13T17:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.160160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.160208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.160217 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.160680 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.160705 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:32Z","lastTransitionTime":"2025-12-13T17:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.263541 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.263592 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.263605 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.263626 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.263641 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:32Z","lastTransitionTime":"2025-12-13T17:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.365188 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.365227 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.365235 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.365248 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.365257 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:32Z","lastTransitionTime":"2025-12-13T17:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.467204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.467238 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.467248 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.467262 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.467272 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:32Z","lastTransitionTime":"2025-12-13T17:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.569784 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.569835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.569843 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.569858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.569868 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:32Z","lastTransitionTime":"2025-12-13T17:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.672206 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.672268 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.672288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.672315 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.672333 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:32Z","lastTransitionTime":"2025-12-13T17:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.774723 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.774765 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.774775 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.774806 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.774815 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:32Z","lastTransitionTime":"2025-12-13T17:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.877415 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.877472 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.877483 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.877497 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.877509 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:32Z","lastTransitionTime":"2025-12-13T17:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.979638 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.979688 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.979700 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.979718 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:32 crc kubenswrapper[4989]: I1213 17:17:32.979729 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:32Z","lastTransitionTime":"2025-12-13T17:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.014157 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:33 crc kubenswrapper[4989]: E1213 17:17:33.014280 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.082430 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.082494 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.082512 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.082537 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.082554 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:33Z","lastTransitionTime":"2025-12-13T17:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.186258 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.186329 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.186352 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.186376 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.186393 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:33Z","lastTransitionTime":"2025-12-13T17:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.288976 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.289013 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.289024 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.289040 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.289052 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:33Z","lastTransitionTime":"2025-12-13T17:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.391189 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.391237 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.391251 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.391270 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.391286 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:33Z","lastTransitionTime":"2025-12-13T17:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.493562 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.493619 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.493633 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.493659 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.493677 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:33Z","lastTransitionTime":"2025-12-13T17:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.596459 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.596509 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.596519 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.596551 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.596560 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:33Z","lastTransitionTime":"2025-12-13T17:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.698579 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.698618 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.698629 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.698644 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.698654 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:33Z","lastTransitionTime":"2025-12-13T17:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.800611 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.800645 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.800653 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.800667 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.800676 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:33Z","lastTransitionTime":"2025-12-13T17:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.903438 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.903543 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.903569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.903604 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:33 crc kubenswrapper[4989]: I1213 17:17:33.903630 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:33Z","lastTransitionTime":"2025-12-13T17:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.006410 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.006503 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.006526 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.006553 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.006573 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:34Z","lastTransitionTime":"2025-12-13T17:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.013742 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.013856 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.013780 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:34 crc kubenswrapper[4989]: E1213 17:17:34.013990 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:34 crc kubenswrapper[4989]: E1213 17:17:34.014116 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:34 crc kubenswrapper[4989]: E1213 17:17:34.014225 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.109569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.109644 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.109669 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.109700 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.109726 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:34Z","lastTransitionTime":"2025-12-13T17:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.211857 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.211897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.211910 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.211926 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.211941 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:34Z","lastTransitionTime":"2025-12-13T17:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.314748 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.314818 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.314833 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.314851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.314864 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:34Z","lastTransitionTime":"2025-12-13T17:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.417347 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.417399 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.417410 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.417429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.417441 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:34Z","lastTransitionTime":"2025-12-13T17:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.519659 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.519705 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.519716 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.519732 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.519744 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:34Z","lastTransitionTime":"2025-12-13T17:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.622083 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.622139 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.622153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.622172 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.622185 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:34Z","lastTransitionTime":"2025-12-13T17:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.724415 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.724475 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.724492 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.724517 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.724535 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:34Z","lastTransitionTime":"2025-12-13T17:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.828448 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.828490 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.828501 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.828519 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.828530 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:34Z","lastTransitionTime":"2025-12-13T17:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.930470 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.930542 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.930553 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.930570 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:34 crc kubenswrapper[4989]: I1213 17:17:34.930583 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:34Z","lastTransitionTime":"2025-12-13T17:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.014002 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:35 crc kubenswrapper[4989]: E1213 17:17:35.014168 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.033518 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.033564 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.033582 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.033606 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.033623 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:35Z","lastTransitionTime":"2025-12-13T17:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.135924 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.135992 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.136018 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.136085 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.136109 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:35Z","lastTransitionTime":"2025-12-13T17:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.237917 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.237961 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.237972 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.237987 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.237998 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:35Z","lastTransitionTime":"2025-12-13T17:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.340521 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.340594 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.340626 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.340736 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.340775 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:35Z","lastTransitionTime":"2025-12-13T17:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.443449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.443489 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.443501 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.443515 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.443525 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:35Z","lastTransitionTime":"2025-12-13T17:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.546169 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.546208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.546219 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.546236 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.546248 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:35Z","lastTransitionTime":"2025-12-13T17:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.648707 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.648740 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.648748 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.648762 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.648771 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:35Z","lastTransitionTime":"2025-12-13T17:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.750815 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.750845 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.750857 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.750871 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.750880 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:35Z","lastTransitionTime":"2025-12-13T17:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.853135 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.853168 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.853176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.853190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.853200 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:35Z","lastTransitionTime":"2025-12-13T17:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.955283 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.955311 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.955319 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.955332 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:35 crc kubenswrapper[4989]: I1213 17:17:35.955341 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:35Z","lastTransitionTime":"2025-12-13T17:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.013536 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.013566 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.013605 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:36 crc kubenswrapper[4989]: E1213 17:17:36.013718 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:36 crc kubenswrapper[4989]: E1213 17:17:36.013811 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:36 crc kubenswrapper[4989]: E1213 17:17:36.013870 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.026265 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.039849 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.056901 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.057319 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.057354 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.057364 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.057381 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.057391 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:36Z","lastTransitionTime":"2025-12-13T17:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.072072 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.085928 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.100523 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.113646 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.126451 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.145712 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:19Z\\\",\\\"message\\\":\\\"r] map[service.beta.openshift.io/serving-cert-secret-name:canary-serving-cert] [{apps/v1 daemonset ingress-canary f5a2759b-dc3c-483d-93f0-055bac962b12 0xc000659f87 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:8443-tcp,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:8888-tcp,Protocol:TCP,Port:8888,TargetPort:{0 8888 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{ingresscanary.operator.openshift.io/daemonset-ingresscanary: canary_controller,},ClusterIP:10.217.5.34,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.34],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1213 17:17:19.137223 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, hand\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.155608 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.159049 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.159103 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.159118 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.159135 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.159147 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:36Z","lastTransitionTime":"2025-12-13T17:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.173540 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.185533 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.195859 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30798eda-9ae1-43fd-918b-a13c4be89568\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.210315 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.221970 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.232720 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.241820 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.253685 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:36Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.261469 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.261497 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.261506 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.261520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.261528 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:36Z","lastTransitionTime":"2025-12-13T17:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.363260 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.363302 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.363313 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.363379 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.363416 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:36Z","lastTransitionTime":"2025-12-13T17:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.465214 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.465244 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.465254 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.465267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.465276 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:36Z","lastTransitionTime":"2025-12-13T17:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.567672 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.567704 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.567712 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.567728 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.567736 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:36Z","lastTransitionTime":"2025-12-13T17:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.670006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.670040 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.670049 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.670064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.670072 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:36Z","lastTransitionTime":"2025-12-13T17:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.771774 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.771837 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.771846 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.771862 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.771871 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:36Z","lastTransitionTime":"2025-12-13T17:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.873849 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.873889 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.873898 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.873912 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.873921 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:36Z","lastTransitionTime":"2025-12-13T17:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.976536 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.976583 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.976593 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.976609 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:36 crc kubenswrapper[4989]: I1213 17:17:36.976620 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:36Z","lastTransitionTime":"2025-12-13T17:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.013558 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:37 crc kubenswrapper[4989]: E1213 17:17:37.013704 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.079110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.079154 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.079165 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.079181 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.079192 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:37Z","lastTransitionTime":"2025-12-13T17:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.181697 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.181746 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.181757 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.181773 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.181801 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:37Z","lastTransitionTime":"2025-12-13T17:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.270094 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:37 crc kubenswrapper[4989]: E1213 17:17:37.270219 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:37 crc kubenswrapper[4989]: E1213 17:17:37.270267 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs podName:7d912915-788e-412a-bae8-6eccd6b4c238 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:09.270255219 +0000 UTC m=+103.876702357 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs") pod "network-metrics-daemon-lfpf8" (UID: "7d912915-788e-412a-bae8-6eccd6b4c238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.284165 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.284198 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.284208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.284225 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.284238 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:37Z","lastTransitionTime":"2025-12-13T17:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.386272 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.386313 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.386325 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.386339 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.386349 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:37Z","lastTransitionTime":"2025-12-13T17:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.488867 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.488936 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.488945 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.488958 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.488969 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:37Z","lastTransitionTime":"2025-12-13T17:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.591695 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.591723 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.591751 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.591764 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.591775 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:37Z","lastTransitionTime":"2025-12-13T17:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.694501 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.694549 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.694561 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.694578 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.694591 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:37Z","lastTransitionTime":"2025-12-13T17:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.797214 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.797251 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.797262 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.797277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.797287 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:37Z","lastTransitionTime":"2025-12-13T17:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.900300 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.900365 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.900382 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.900400 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:37 crc kubenswrapper[4989]: I1213 17:17:37.900435 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:37Z","lastTransitionTime":"2025-12-13T17:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.002612 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.002644 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.002653 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.002667 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.002676 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.016073 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:38 crc kubenswrapper[4989]: E1213 17:17:38.016156 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.016273 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:38 crc kubenswrapper[4989]: E1213 17:17:38.016322 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.016413 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:38 crc kubenswrapper[4989]: E1213 17:17:38.016459 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.105105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.105148 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.105161 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.105180 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.105192 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.207475 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.207510 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.207520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.207534 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.207545 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.309489 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.309524 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.309533 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.309547 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.309557 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.413867 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.413907 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.413916 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.413931 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.413942 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.415183 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hllvq_1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf/kube-multus/0.log" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.415231 4989 generic.go:334] "Generic (PLEG): container finished" podID="1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf" containerID="d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494" exitCode=1 Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.415262 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hllvq" event={"ID":"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf","Type":"ContainerDied","Data":"d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.415655 4989 scope.go:117] "RemoveContainer" containerID="d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.428660 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.440652 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30798eda-9ae1-43fd-918b-a13c4be89568\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.454476 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.468603 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:37Z\\\",\\\"message\\\":\\\"2025-12-13T17:16:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8\\\\n2025-12-13T17:16:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8 to /host/opt/cni/bin/\\\\n2025-12-13T17:16:52Z [verbose] multus-daemon started\\\\n2025-12-13T17:16:52Z [verbose] Readiness Indicator file check\\\\n2025-12-13T17:17:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.479692 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.491768 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.505809 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.506137 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.506165 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.506176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.506191 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.506200 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: E1213 17:17:38.521723 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.525497 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.525564 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.525580 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.525597 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.525608 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.527007 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.539508 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: E1213 17:17:38.539895 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.542393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.542500 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.542594 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.542675 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.542761 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: E1213 17:17:38.556359 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.557567 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.563864 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.563925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.563936 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.563953 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.563965 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.571182 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: E1213 17:17:38.574776 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.582240 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.582276 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.582285 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.582301 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.582311 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.591598 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: E1213 17:17:38.593015 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: E1213 17:17:38.593165 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.594768 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.594812 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.594822 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.594833 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.594842 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.601810 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.612917 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.623094 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.639204 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:19Z\\\",\\\"message\\\":\\\"r] map[service.beta.openshift.io/serving-cert-secret-name:canary-serving-cert] [{apps/v1 daemonset ingress-canary f5a2759b-dc3c-483d-93f0-055bac962b12 0xc000659f87 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:8443-tcp,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:8888-tcp,Protocol:TCP,Port:8888,TargetPort:{0 8888 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{ingresscanary.operator.openshift.io/daemonset-ingresscanary: canary_controller,},ClusterIP:10.217.5.34,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.34],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1213 17:17:19.137223 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, hand\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.650227 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.661366 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:38Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.696652 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.696680 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.696689 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.696702 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.696711 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.799305 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.799338 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.799348 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.799363 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.799373 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.901469 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.901504 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.901514 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.901530 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:38 crc kubenswrapper[4989]: I1213 17:17:38.901540 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:38Z","lastTransitionTime":"2025-12-13T17:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.004162 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.004200 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.004210 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.004224 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.004236 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:39Z","lastTransitionTime":"2025-12-13T17:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.014435 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:39 crc kubenswrapper[4989]: E1213 17:17:39.014546 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.106941 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.106986 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.106997 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.107015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.107027 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:39Z","lastTransitionTime":"2025-12-13T17:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.209098 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.209146 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.209156 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.209171 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.209180 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:39Z","lastTransitionTime":"2025-12-13T17:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.311208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.311245 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.311254 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.311267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.311276 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:39Z","lastTransitionTime":"2025-12-13T17:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.413772 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.413833 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.413844 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.413859 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.413869 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:39Z","lastTransitionTime":"2025-12-13T17:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.418992 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hllvq_1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf/kube-multus/0.log" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.419035 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hllvq" event={"ID":"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf","Type":"ContainerStarted","Data":"4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c"} Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.431743 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.444049 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.456691 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.469404 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.477390 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.494429 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.506866 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.516252 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.516297 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.516308 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.516326 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.516339 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:39Z","lastTransitionTime":"2025-12-13T17:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.519102 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.531971 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.557705 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:19Z\\\",\\\"message\\\":\\\"r] map[service.beta.openshift.io/serving-cert-secret-name:canary-serving-cert] [{apps/v1 daemonset ingress-canary f5a2759b-dc3c-483d-93f0-055bac962b12 0xc000659f87 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:8443-tcp,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:8888-tcp,Protocol:TCP,Port:8888,TargetPort:{0 8888 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{ingresscanary.operator.openshift.io/daemonset-ingresscanary: canary_controller,},ClusterIP:10.217.5.34,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.34],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1213 17:17:19.137223 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, hand\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.573021 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.587380 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.608224 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.618496 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.618538 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.618550 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.618569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.618581 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:39Z","lastTransitionTime":"2025-12-13T17:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.624959 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30798eda-9ae1-43fd-918b-a13c4be89568\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.640877 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.657334 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:37Z\\\",\\\"message\\\":\\\"2025-12-13T17:16:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8\\\\n2025-12-13T17:16:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8 to /host/opt/cni/bin/\\\\n2025-12-13T17:16:52Z [verbose] multus-daemon started\\\\n2025-12-13T17:16:52Z [verbose] Readiness Indicator file check\\\\n2025-12-13T17:17:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.669278 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.681495 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:39Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.720265 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.720308 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.720321 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.720337 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.720348 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:39Z","lastTransitionTime":"2025-12-13T17:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.822835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.822927 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.822954 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.822980 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.822998 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:39Z","lastTransitionTime":"2025-12-13T17:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.926869 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.926934 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.926947 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.926967 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:39 crc kubenswrapper[4989]: I1213 17:17:39.926985 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:39Z","lastTransitionTime":"2025-12-13T17:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.013571 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.013602 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:40 crc kubenswrapper[4989]: E1213 17:17:40.013692 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:40 crc kubenswrapper[4989]: E1213 17:17:40.013778 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.013802 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:40 crc kubenswrapper[4989]: E1213 17:17:40.013849 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.028784 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.028825 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.028833 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.028844 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.028854 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:40Z","lastTransitionTime":"2025-12-13T17:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.131372 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.131407 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.131415 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.131431 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.131439 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:40Z","lastTransitionTime":"2025-12-13T17:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.234060 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.234099 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.234109 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.234128 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.234137 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:40Z","lastTransitionTime":"2025-12-13T17:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.336371 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.336415 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.336426 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.336440 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.336453 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:40Z","lastTransitionTime":"2025-12-13T17:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.438178 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.438215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.438223 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.438236 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.438245 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:40Z","lastTransitionTime":"2025-12-13T17:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.540858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.540890 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.540900 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.540913 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.540922 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:40Z","lastTransitionTime":"2025-12-13T17:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.643195 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.643234 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.643243 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.643256 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.643278 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:40Z","lastTransitionTime":"2025-12-13T17:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.745520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.745574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.745586 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.745611 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.745624 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:40Z","lastTransitionTime":"2025-12-13T17:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.848318 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.848355 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.848363 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.848378 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.848388 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:40Z","lastTransitionTime":"2025-12-13T17:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.951548 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.951599 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.951610 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.951630 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:40 crc kubenswrapper[4989]: I1213 17:17:40.951645 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:40Z","lastTransitionTime":"2025-12-13T17:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.014348 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:41 crc kubenswrapper[4989]: E1213 17:17:41.014480 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.015154 4989 scope.go:117] "RemoveContainer" containerID="836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.054505 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.054813 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.054842 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.054861 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.054874 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:41Z","lastTransitionTime":"2025-12-13T17:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.157647 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.157676 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.157685 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.157699 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.157708 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:41Z","lastTransitionTime":"2025-12-13T17:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.260317 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.260349 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.260358 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.260372 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.260380 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:41Z","lastTransitionTime":"2025-12-13T17:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.362770 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.362855 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.362872 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.362897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.362914 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:41Z","lastTransitionTime":"2025-12-13T17:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.426232 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/2.log" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.428871 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1"} Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.429543 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.441342 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.454766 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.464642 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.464678 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.464689 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.464706 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.464714 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:41Z","lastTransitionTime":"2025-12-13T17:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.468268 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30798eda-9ae1-43fd-918b-a13c4be89568\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.479020 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.489796 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:37Z\\\",\\\"message\\\":\\\"2025-12-13T17:16:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8\\\\n2025-12-13T17:16:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8 to /host/opt/cni/bin/\\\\n2025-12-13T17:16:52Z [verbose] multus-daemon started\\\\n2025-12-13T17:16:52Z [verbose] Readiness Indicator file check\\\\n2025-12-13T17:17:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.497782 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.506738 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.519717 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.538229 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.550313 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.566604 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.566654 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.566666 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.566684 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.566697 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:41Z","lastTransitionTime":"2025-12-13T17:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.570593 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.582842 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.599447 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.610892 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.623446 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.632011 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.648114 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:19Z\\\",\\\"message\\\":\\\"r] map[service.beta.openshift.io/serving-cert-secret-name:canary-serving-cert] [{apps/v1 daemonset ingress-canary f5a2759b-dc3c-483d-93f0-055bac962b12 0xc000659f87 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:8443-tcp,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:8888-tcp,Protocol:TCP,Port:8888,TargetPort:{0 8888 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{ingresscanary.operator.openshift.io/daemonset-ingresscanary: canary_controller,},ClusterIP:10.217.5.34,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.34],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1213 17:17:19.137223 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, hand\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.655922 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:41Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.668640 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.668674 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.668685 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.668699 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.668708 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:41Z","lastTransitionTime":"2025-12-13T17:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.770914 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.770973 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.770989 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.771008 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.771022 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:41Z","lastTransitionTime":"2025-12-13T17:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.872755 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.872804 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.872813 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.872827 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.872837 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:41Z","lastTransitionTime":"2025-12-13T17:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.975941 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.975981 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.975990 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.976006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:41 crc kubenswrapper[4989]: I1213 17:17:41.976018 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:41Z","lastTransitionTime":"2025-12-13T17:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.013974 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.013991 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.014110 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:42 crc kubenswrapper[4989]: E1213 17:17:42.014232 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:42 crc kubenswrapper[4989]: E1213 17:17:42.014329 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:42 crc kubenswrapper[4989]: E1213 17:17:42.014408 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.078394 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.078436 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.078446 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.078463 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.078473 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:42Z","lastTransitionTime":"2025-12-13T17:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.181209 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.181253 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.181264 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.181282 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.181295 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:42Z","lastTransitionTime":"2025-12-13T17:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.284414 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.284455 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.284466 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.284482 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.284495 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:42Z","lastTransitionTime":"2025-12-13T17:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.386651 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.386684 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.386692 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.386705 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.386715 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:42Z","lastTransitionTime":"2025-12-13T17:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.435161 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/3.log" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.436486 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/2.log" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.440388 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1" exitCode=1 Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.440471 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1"} Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.440560 4989 scope.go:117] "RemoveContainer" containerID="836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.441548 4989 scope.go:117] "RemoveContainer" containerID="8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1" Dec 13 17:17:42 crc kubenswrapper[4989]: E1213 17:17:42.441905 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.468043 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.482553 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.488914 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.488965 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.488982 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.489006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.489023 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:42Z","lastTransitionTime":"2025-12-13T17:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.495560 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.509765 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.533125 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836de747acd5719d16270292a4ab19aa62359334af15a25e3990d336fdd9a9ea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:19Z\\\",\\\"message\\\":\\\"r] map[service.beta.openshift.io/serving-cert-secret-name:canary-serving-cert] [{apps/v1 daemonset ingress-canary f5a2759b-dc3c-483d-93f0-055bac962b12 0xc000659f87 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:8443-tcp,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},ServicePort{Name:8888-tcp,Protocol:TCP,Port:8888,TargetPort:{0 8888 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{ingresscanary.operator.openshift.io/daemonset-ingresscanary: canary_controller,},ClusterIP:10.217.5.34,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.34],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1213 17:17:19.137223 6628 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, hand\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:41Z\\\",\\\"message\\\":\\\"Family:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1213 17:17:41.791836 6994 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1213 17:17:41.791993 6994 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.548356 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.566406 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.582257 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.591410 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.591452 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.591467 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.591485 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.591498 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:42Z","lastTransitionTime":"2025-12-13T17:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.595229 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30798eda-9ae1-43fd-918b-a13c4be89568\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.609990 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.629473 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:37Z\\\",\\\"message\\\":\\\"2025-12-13T17:16:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8\\\\n2025-12-13T17:16:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8 to /host/opt/cni/bin/\\\\n2025-12-13T17:16:52Z [verbose] multus-daemon started\\\\n2025-12-13T17:16:52Z [verbose] Readiness Indicator file check\\\\n2025-12-13T17:17:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.642187 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.654047 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.671416 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.692214 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.694248 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.694279 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.694290 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.694307 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.694319 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:42Z","lastTransitionTime":"2025-12-13T17:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.707125 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.724076 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.732959 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:42Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.797183 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.797229 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.797242 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.797257 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.797269 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:42Z","lastTransitionTime":"2025-12-13T17:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.900286 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.900322 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.900331 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.900346 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:42 crc kubenswrapper[4989]: I1213 17:17:42.900356 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:42Z","lastTransitionTime":"2025-12-13T17:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.002219 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.002269 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.002286 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.002309 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.002329 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:43Z","lastTransitionTime":"2025-12-13T17:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.013734 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:43 crc kubenswrapper[4989]: E1213 17:17:43.013939 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.104523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.104589 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.104610 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.104688 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.104716 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:43Z","lastTransitionTime":"2025-12-13T17:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.207690 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.207776 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.207829 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.207859 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.207881 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:43Z","lastTransitionTime":"2025-12-13T17:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.310311 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.310353 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.310362 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.310379 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.310393 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:43Z","lastTransitionTime":"2025-12-13T17:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.413111 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.413149 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.413160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.413176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.413188 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:43Z","lastTransitionTime":"2025-12-13T17:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.444882 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/3.log" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.449329 4989 scope.go:117] "RemoveContainer" containerID="8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1" Dec 13 17:17:43 crc kubenswrapper[4989]: E1213 17:17:43.449642 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.467776 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.480050 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:37Z\\\",\\\"message\\\":\\\"2025-12-13T17:16:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8\\\\n2025-12-13T17:16:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8 to /host/opt/cni/bin/\\\\n2025-12-13T17:16:52Z [verbose] multus-daemon started\\\\n2025-12-13T17:16:52Z [verbose] Readiness Indicator file check\\\\n2025-12-13T17:17:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.492110 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.504170 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.515663 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.515722 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.515740 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.515766 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.515783 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:43Z","lastTransitionTime":"2025-12-13T17:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.520140 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.535184 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30798eda-9ae1-43fd-918b-a13c4be89568\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.547446 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.564105 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.584209 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.595777 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.614521 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.618326 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.618366 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.618375 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.618387 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.618396 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:43Z","lastTransitionTime":"2025-12-13T17:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.633278 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.650455 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.672815 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:41Z\\\",\\\"message\\\":\\\"Family:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1213 17:17:41.791836 6994 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1213 17:17:41.791993 6994 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.685426 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.708474 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.723844 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.723897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.723909 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.723925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.723939 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:43Z","lastTransitionTime":"2025-12-13T17:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.729695 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.745535 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:43Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.825972 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.826003 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.826011 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.826025 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.826034 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:43Z","lastTransitionTime":"2025-12-13T17:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.928540 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.928576 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.928586 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.928601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:43 crc kubenswrapper[4989]: I1213 17:17:43.928612 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:43Z","lastTransitionTime":"2025-12-13T17:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.013943 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.014222 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:44 crc kubenswrapper[4989]: E1213 17:17:44.014204 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.014244 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:44 crc kubenswrapper[4989]: E1213 17:17:44.014358 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:44 crc kubenswrapper[4989]: E1213 17:17:44.014420 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.025782 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.032335 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.032412 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.032440 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.032481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.032506 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:44Z","lastTransitionTime":"2025-12-13T17:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.134942 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.135049 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.135069 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.135096 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.135116 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:44Z","lastTransitionTime":"2025-12-13T17:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.238246 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.238289 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.238303 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.238318 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.238331 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:44Z","lastTransitionTime":"2025-12-13T17:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.340470 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.340533 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.340553 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.340577 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.340596 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:44Z","lastTransitionTime":"2025-12-13T17:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.443660 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.443741 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.443767 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.443865 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.443895 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:44Z","lastTransitionTime":"2025-12-13T17:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.545771 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.545877 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.545898 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.545927 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.545948 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:44Z","lastTransitionTime":"2025-12-13T17:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.648436 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.648488 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.648504 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.648525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.648542 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:44Z","lastTransitionTime":"2025-12-13T17:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.751427 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.751462 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.751471 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.751486 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.751496 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:44Z","lastTransitionTime":"2025-12-13T17:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.854376 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.854443 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.854467 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.854493 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.854513 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:44Z","lastTransitionTime":"2025-12-13T17:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.957350 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.957423 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.957444 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.957475 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:44 crc kubenswrapper[4989]: I1213 17:17:44.957497 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:44Z","lastTransitionTime":"2025-12-13T17:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.014137 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:45 crc kubenswrapper[4989]: E1213 17:17:45.014271 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.059919 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.059947 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.059957 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.059971 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.059982 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:45Z","lastTransitionTime":"2025-12-13T17:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.162130 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.162169 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.162178 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.162191 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.162202 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:45Z","lastTransitionTime":"2025-12-13T17:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.265247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.265280 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.265289 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.265303 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.265313 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:45Z","lastTransitionTime":"2025-12-13T17:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.368255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.368301 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.368310 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.368326 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.368338 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:45Z","lastTransitionTime":"2025-12-13T17:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.471253 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.471335 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.471355 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.471382 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.471402 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:45Z","lastTransitionTime":"2025-12-13T17:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.575606 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.575705 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.575727 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.575756 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.575775 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:45Z","lastTransitionTime":"2025-12-13T17:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.678723 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.678768 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.678778 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.678808 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.678820 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:45Z","lastTransitionTime":"2025-12-13T17:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.782224 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.782296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.782313 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.782343 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.782362 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:45Z","lastTransitionTime":"2025-12-13T17:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.885749 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.885891 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.885917 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.885954 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.885981 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:45Z","lastTransitionTime":"2025-12-13T17:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.989273 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.989339 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.989350 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.989373 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:45 crc kubenswrapper[4989]: I1213 17:17:45.989385 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:45Z","lastTransitionTime":"2025-12-13T17:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.014155 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.014190 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:46 crc kubenswrapper[4989]: E1213 17:17:46.014356 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.014466 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:46 crc kubenswrapper[4989]: E1213 17:17:46.014707 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:46 crc kubenswrapper[4989]: E1213 17:17:46.014985 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.042189 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.060397 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.078106 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"390325ff-bbf3-4c16-88c7-cdb04976409a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961b7ddbae6e098a3f15adc1253a50fc94fcd98c362f248e976570a9a391769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386f0a89a932893d541d90c3daec49677e1ae574c5bc75138187d7f67b0b0c9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386f0a89a932893d541d90c3daec49677e1ae574c5bc75138187d7f67b0b0c9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.093523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.093609 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.093636 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.093673 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.093706 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:46Z","lastTransitionTime":"2025-12-13T17:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.102471 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.125377 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.143381 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.174299 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:41Z\\\",\\\"message\\\":\\\"Family:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1213 17:17:41.791836 6994 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1213 17:17:41.791993 6994 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.190641 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.196857 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.196909 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.196927 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.196950 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.196968 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:46Z","lastTransitionTime":"2025-12-13T17:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.220695 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.236061 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.254876 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.269336 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.283593 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.298121 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.300668 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.300768 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.300835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.300945 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.300965 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:46Z","lastTransitionTime":"2025-12-13T17:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.316837 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.333619 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.352252 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30798eda-9ae1-43fd-918b-a13c4be89568\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.368223 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.387014 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:37Z\\\",\\\"message\\\":\\\"2025-12-13T17:16:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8\\\\n2025-12-13T17:16:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8 to /host/opt/cni/bin/\\\\n2025-12-13T17:16:52Z [verbose] multus-daemon started\\\\n2025-12-13T17:16:52Z [verbose] Readiness Indicator file check\\\\n2025-12-13T17:17:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:46Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.402839 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.402888 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.402896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.402909 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.402917 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:46Z","lastTransitionTime":"2025-12-13T17:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.505881 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.505926 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.505938 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.505951 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.505960 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:46Z","lastTransitionTime":"2025-12-13T17:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.608825 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.608900 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.608919 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.608955 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.608979 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:46Z","lastTransitionTime":"2025-12-13T17:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.712050 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.712144 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.712166 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.712200 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.712221 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:46Z","lastTransitionTime":"2025-12-13T17:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.815774 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.815865 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.815880 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.815904 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.815919 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:46Z","lastTransitionTime":"2025-12-13T17:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.919009 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.919064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.919077 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.919106 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:46 crc kubenswrapper[4989]: I1213 17:17:46.919123 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:46Z","lastTransitionTime":"2025-12-13T17:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.014220 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:47 crc kubenswrapper[4989]: E1213 17:17:47.014354 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.022154 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.022180 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.022190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.022202 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.022210 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:47Z","lastTransitionTime":"2025-12-13T17:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.125757 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.125802 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.125813 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.125828 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.125838 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:47Z","lastTransitionTime":"2025-12-13T17:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.229319 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.229361 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.229371 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.229388 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.229397 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:47Z","lastTransitionTime":"2025-12-13T17:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.333519 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.333612 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.333633 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.333666 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.333687 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:47Z","lastTransitionTime":"2025-12-13T17:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.436945 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.436997 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.437009 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.437030 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.437043 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:47Z","lastTransitionTime":"2025-12-13T17:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.538819 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.538851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.538859 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.538873 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.538881 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:47Z","lastTransitionTime":"2025-12-13T17:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.641022 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.641056 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.641066 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.641081 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.641091 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:47Z","lastTransitionTime":"2025-12-13T17:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.744080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.744116 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.744125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.744138 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.744147 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:47Z","lastTransitionTime":"2025-12-13T17:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.846724 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.846764 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.846774 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.846842 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.846857 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:47Z","lastTransitionTime":"2025-12-13T17:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.897834 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:17:47 crc kubenswrapper[4989]: E1213 17:17:47.898021 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.897996781 +0000 UTC m=+146.504443919 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.949313 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.949353 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.949364 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.949376 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.949386 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:47Z","lastTransitionTime":"2025-12-13T17:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:47 crc kubenswrapper[4989]: E1213 17:17:47.998854 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:17:47 crc kubenswrapper[4989]: E1213 17:17:47.998974 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.998945101 +0000 UTC m=+146.605392279 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.998672 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:47 crc kubenswrapper[4989]: I1213 17:17:47.999318 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:47 crc kubenswrapper[4989]: E1213 17:17:47.999513 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:17:47 crc kubenswrapper[4989]: E1213 17:17:47.999668 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:17:47 crc kubenswrapper[4989]: E1213 17:17:47.999687 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:17:47 crc kubenswrapper[4989]: E1213 17:17:47.999739 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.999723474 +0000 UTC m=+146.606170652 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:17:47 crc kubenswrapper[4989]: E1213 17:17:47.999825 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:17:47 crc kubenswrapper[4989]: E1213 17:17:47.999869 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.999856588 +0000 UTC m=+146.606303726 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.000211 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.000330 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.000503 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.000601 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.000623 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.000689 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:52.000667462 +0000 UTC m=+146.607114670 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.013608 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.013637 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.013707 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.013872 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.014167 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.014300 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.051219 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.051249 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.051261 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.051275 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.051285 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.153747 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.153866 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.153893 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.153921 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.153943 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.256600 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.256657 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.256668 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.256700 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.256712 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.358680 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.358740 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.358754 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.358768 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.358778 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.462011 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.462107 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.462129 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.462159 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.462182 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.564966 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.565001 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.565010 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.565024 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.565033 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.668143 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.668452 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.668466 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.668481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.668491 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.716749 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.716815 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.716856 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.716876 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.716906 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.732082 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:48Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.739276 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.739391 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.739410 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.739435 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.739452 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.757918 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:48Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.762760 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.762812 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.762821 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.762835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.762844 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.774580 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:48Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.778366 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.778396 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.778406 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.778418 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.778427 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.789510 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:48Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.792939 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.792971 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.792979 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.792991 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.793001 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.804364 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:48Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:48 crc kubenswrapper[4989]: E1213 17:17:48.804523 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.805729 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.805763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.805775 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.805807 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.805820 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.909183 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.909242 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.909254 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.909273 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:48 crc kubenswrapper[4989]: I1213 17:17:48.909286 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:48Z","lastTransitionTime":"2025-12-13T17:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.012000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.012058 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.012078 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.012101 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.012116 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:49Z","lastTransitionTime":"2025-12-13T17:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.013584 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:49 crc kubenswrapper[4989]: E1213 17:17:49.013781 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.115156 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.115201 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.115242 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.115264 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.115276 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:49Z","lastTransitionTime":"2025-12-13T17:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.217867 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.217963 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.217982 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.218063 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.218084 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:49Z","lastTransitionTime":"2025-12-13T17:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.320333 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.320448 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.320563 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.320636 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.320741 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:49Z","lastTransitionTime":"2025-12-13T17:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.422769 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.422843 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.422855 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.422873 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.422886 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:49Z","lastTransitionTime":"2025-12-13T17:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.525809 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.525848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.525857 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.525872 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.525882 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:49Z","lastTransitionTime":"2025-12-13T17:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.628530 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.628576 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.628591 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.628612 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.628634 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:49Z","lastTransitionTime":"2025-12-13T17:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.730776 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.730868 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.730878 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.730893 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.730902 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:49Z","lastTransitionTime":"2025-12-13T17:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.833690 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.833739 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.833755 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.833772 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.833808 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:49Z","lastTransitionTime":"2025-12-13T17:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.937286 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.937317 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.937329 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.937344 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:49 crc kubenswrapper[4989]: I1213 17:17:49.937355 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:49Z","lastTransitionTime":"2025-12-13T17:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.014391 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:50 crc kubenswrapper[4989]: E1213 17:17:50.014507 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.014658 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:50 crc kubenswrapper[4989]: E1213 17:17:50.014698 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.014832 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:50 crc kubenswrapper[4989]: E1213 17:17:50.014896 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.039725 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.039759 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.039769 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.039980 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.039994 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:50Z","lastTransitionTime":"2025-12-13T17:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.142005 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.142060 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.142073 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.142084 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.142126 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:50Z","lastTransitionTime":"2025-12-13T17:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.244565 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.244597 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.244605 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.244617 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.244627 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:50Z","lastTransitionTime":"2025-12-13T17:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.346442 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.346539 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.346563 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.346590 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.346610 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:50Z","lastTransitionTime":"2025-12-13T17:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.448835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.448868 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.448878 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.448892 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.448904 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:50Z","lastTransitionTime":"2025-12-13T17:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.554580 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.554643 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.554665 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.554686 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.554699 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:50Z","lastTransitionTime":"2025-12-13T17:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.657556 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.657600 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.657616 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.657637 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.657654 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:50Z","lastTransitionTime":"2025-12-13T17:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.760505 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.760574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.760593 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.760617 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.760634 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:50Z","lastTransitionTime":"2025-12-13T17:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.863188 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.863228 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.863239 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.863256 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.863267 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:50Z","lastTransitionTime":"2025-12-13T17:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.965346 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.965389 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.965401 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.965417 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:50 crc kubenswrapper[4989]: I1213 17:17:50.965430 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:50Z","lastTransitionTime":"2025-12-13T17:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.014098 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:51 crc kubenswrapper[4989]: E1213 17:17:51.014240 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.068142 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.068197 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.068215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.068238 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.068255 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:51Z","lastTransitionTime":"2025-12-13T17:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.171130 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.171186 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.171196 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.171215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.171228 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:51Z","lastTransitionTime":"2025-12-13T17:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.273678 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.273705 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.273713 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.273727 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.273736 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:51Z","lastTransitionTime":"2025-12-13T17:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.376153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.376211 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.376226 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.376249 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.376265 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:51Z","lastTransitionTime":"2025-12-13T17:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.477913 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.477952 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.477968 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.477985 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.477997 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:51Z","lastTransitionTime":"2025-12-13T17:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.580441 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.580544 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.580557 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.580574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.580585 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:51Z","lastTransitionTime":"2025-12-13T17:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.682403 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.682455 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.682465 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.682482 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.682492 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:51Z","lastTransitionTime":"2025-12-13T17:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.788063 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.788098 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.788109 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.788123 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.788133 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:51Z","lastTransitionTime":"2025-12-13T17:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.891294 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.891351 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.891370 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.891392 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.891410 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:51Z","lastTransitionTime":"2025-12-13T17:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.994356 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.994387 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.994400 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.994416 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:51 crc kubenswrapper[4989]: I1213 17:17:51.994427 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:51Z","lastTransitionTime":"2025-12-13T17:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.014263 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.014286 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.014441 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:52 crc kubenswrapper[4989]: E1213 17:17:52.014536 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:52 crc kubenswrapper[4989]: E1213 17:17:52.014718 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:52 crc kubenswrapper[4989]: E1213 17:17:52.014741 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.097212 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.097247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.097255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.097269 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.097278 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:52Z","lastTransitionTime":"2025-12-13T17:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.200321 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.200387 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.200405 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.200429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.200447 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:52Z","lastTransitionTime":"2025-12-13T17:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.303017 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.303324 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.303332 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.303348 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.303359 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:52Z","lastTransitionTime":"2025-12-13T17:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.406396 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.406473 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.406499 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.406529 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.406552 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:52Z","lastTransitionTime":"2025-12-13T17:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.509776 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.509880 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.509895 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.509925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.509948 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:52Z","lastTransitionTime":"2025-12-13T17:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.612865 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.612950 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.612968 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.612996 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.613015 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:52Z","lastTransitionTime":"2025-12-13T17:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.715998 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.716072 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.716091 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.716117 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.716131 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:52Z","lastTransitionTime":"2025-12-13T17:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.818834 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.818893 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.818907 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.818930 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.818947 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:52Z","lastTransitionTime":"2025-12-13T17:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.920995 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.921044 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.921057 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.921074 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:52 crc kubenswrapper[4989]: I1213 17:17:52.921087 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:52Z","lastTransitionTime":"2025-12-13T17:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.014184 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:53 crc kubenswrapper[4989]: E1213 17:17:53.014426 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.023108 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.023169 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.023191 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.023222 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.023248 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:53Z","lastTransitionTime":"2025-12-13T17:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.125610 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.125655 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.125670 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.125688 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.125706 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:53Z","lastTransitionTime":"2025-12-13T17:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.229083 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.229142 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.229161 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.229185 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.229202 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:53Z","lastTransitionTime":"2025-12-13T17:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.332429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.332494 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.332502 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.332517 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.332526 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:53Z","lastTransitionTime":"2025-12-13T17:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.435770 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.436268 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.436492 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.436658 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.436888 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:53Z","lastTransitionTime":"2025-12-13T17:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.539485 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.539523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.539532 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.539547 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.539557 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:53Z","lastTransitionTime":"2025-12-13T17:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.641905 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.641932 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.641940 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.641953 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.641962 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:53Z","lastTransitionTime":"2025-12-13T17:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.744361 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.744404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.744412 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.744426 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.744435 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:53Z","lastTransitionTime":"2025-12-13T17:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.847143 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.847186 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.847197 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.847212 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.847224 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:53Z","lastTransitionTime":"2025-12-13T17:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.949492 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.949546 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.949557 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.949572 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:53 crc kubenswrapper[4989]: I1213 17:17:53.949584 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:53Z","lastTransitionTime":"2025-12-13T17:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.013617 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.013639 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.013708 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:54 crc kubenswrapper[4989]: E1213 17:17:54.013774 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:54 crc kubenswrapper[4989]: E1213 17:17:54.013904 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:54 crc kubenswrapper[4989]: E1213 17:17:54.014003 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.052090 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.052118 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.052126 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.052138 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.052147 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:54Z","lastTransitionTime":"2025-12-13T17:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.154373 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.154405 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.154415 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.154428 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.154438 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:54Z","lastTransitionTime":"2025-12-13T17:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.256568 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.256698 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.256726 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.256755 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.256777 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:54Z","lastTransitionTime":"2025-12-13T17:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.359104 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.359132 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.359140 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.359153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.359162 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:54Z","lastTransitionTime":"2025-12-13T17:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.462344 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.462490 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.462520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.462597 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.462626 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:54Z","lastTransitionTime":"2025-12-13T17:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.565557 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.565592 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.565605 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.565629 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.565640 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:54Z","lastTransitionTime":"2025-12-13T17:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.667742 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.667777 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.667802 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.667818 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.667828 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:54Z","lastTransitionTime":"2025-12-13T17:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.770210 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.770241 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.770250 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.770263 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.770274 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:54Z","lastTransitionTime":"2025-12-13T17:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.873159 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.873197 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.873206 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.873219 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.873231 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:54Z","lastTransitionTime":"2025-12-13T17:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.976023 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.976089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.976107 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.976131 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:54 crc kubenswrapper[4989]: I1213 17:17:54.976149 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:54Z","lastTransitionTime":"2025-12-13T17:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.014106 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:55 crc kubenswrapper[4989]: E1213 17:17:55.014495 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.014679 4989 scope.go:117] "RemoveContainer" containerID="8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1" Dec 13 17:17:55 crc kubenswrapper[4989]: E1213 17:17:55.014855 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.078279 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.078338 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.078357 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.078380 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.078399 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:55Z","lastTransitionTime":"2025-12-13T17:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.181180 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.181267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.181290 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.181315 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.181335 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:55Z","lastTransitionTime":"2025-12-13T17:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.284021 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.284125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.284144 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.284208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.284682 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:55Z","lastTransitionTime":"2025-12-13T17:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.386849 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.386915 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.386939 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.386969 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.386987 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:55Z","lastTransitionTime":"2025-12-13T17:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.489502 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.489578 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.489604 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.489632 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.489653 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:55Z","lastTransitionTime":"2025-12-13T17:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.594073 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.594145 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.594163 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.594189 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.594207 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:55Z","lastTransitionTime":"2025-12-13T17:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.697321 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.697394 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.697436 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.697472 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.697495 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:55Z","lastTransitionTime":"2025-12-13T17:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.800548 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.800591 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.800606 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.800628 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.800644 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:55Z","lastTransitionTime":"2025-12-13T17:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.904074 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.904137 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.904158 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.904181 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:55 crc kubenswrapper[4989]: I1213 17:17:55.904198 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:55Z","lastTransitionTime":"2025-12-13T17:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.007115 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.007170 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.007188 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.007213 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.007230 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:56Z","lastTransitionTime":"2025-12-13T17:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.014008 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.014021 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:56 crc kubenswrapper[4989]: E1213 17:17:56.014247 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.014030 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:56 crc kubenswrapper[4989]: E1213 17:17:56.014604 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:56 crc kubenswrapper[4989]: E1213 17:17:56.014404 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.031310 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"390325ff-bbf3-4c16-88c7-cdb04976409a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961b7ddbae6e098a3f15adc1253a50fc94fcd98c362f248e976570a9a391769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386f0a89a932893d541d90c3daec49677e1ae574c5bc75138187d7f67b0b0c9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386f0a89a932893d541d90c3daec49677e1ae574c5bc75138187d7f67b0b0c9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.051128 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.072561 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.090547 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.109218 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.109260 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.109272 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.109288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.109299 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:56Z","lastTransitionTime":"2025-12-13T17:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.113505 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.127600 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.150990 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.165488 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.177694 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.193250 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.211640 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.211693 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.211710 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.211734 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.211751 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:56Z","lastTransitionTime":"2025-12-13T17:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.221891 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:41Z\\\",\\\"message\\\":\\\"Family:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1213 17:17:41.791836 6994 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1213 17:17:41.791993 6994 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.233263 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.245989 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.256260 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.269194 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30798eda-9ae1-43fd-918b-a13c4be89568\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.282132 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.293852 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:37Z\\\",\\\"message\\\":\\\"2025-12-13T17:16:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8\\\\n2025-12-13T17:16:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8 to /host/opt/cni/bin/\\\\n2025-12-13T17:16:52Z [verbose] multus-daemon started\\\\n2025-12-13T17:16:52Z [verbose] Readiness Indicator file check\\\\n2025-12-13T17:17:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.306560 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.313964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.314047 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.314060 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.314075 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.314087 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:56Z","lastTransitionTime":"2025-12-13T17:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.317491 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:56Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.416059 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.416086 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.416094 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.416107 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.416116 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:56Z","lastTransitionTime":"2025-12-13T17:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.517698 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.517728 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.517736 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.517750 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.517758 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:56Z","lastTransitionTime":"2025-12-13T17:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.619667 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.619695 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.619703 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.619715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.619723 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:56Z","lastTransitionTime":"2025-12-13T17:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.721520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.721573 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.721591 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.721615 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.721632 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:56Z","lastTransitionTime":"2025-12-13T17:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.824130 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.824173 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.824184 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.824200 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.824210 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:56Z","lastTransitionTime":"2025-12-13T17:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.927316 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.927398 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.927416 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.927446 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:56 crc kubenswrapper[4989]: I1213 17:17:56.927469 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:56Z","lastTransitionTime":"2025-12-13T17:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.014403 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:57 crc kubenswrapper[4989]: E1213 17:17:57.014537 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.030000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.030084 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.030100 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.030147 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.030164 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:57Z","lastTransitionTime":"2025-12-13T17:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.132631 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.132690 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.132710 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.132737 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.132755 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:57Z","lastTransitionTime":"2025-12-13T17:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.235758 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.235818 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.235830 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.235848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.235860 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:57Z","lastTransitionTime":"2025-12-13T17:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.338395 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.338471 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.338501 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.338531 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.338553 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:57Z","lastTransitionTime":"2025-12-13T17:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.441810 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.441848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.441856 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.441886 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.441900 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:57Z","lastTransitionTime":"2025-12-13T17:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.544930 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.545000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.545023 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.545053 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.545075 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:57Z","lastTransitionTime":"2025-12-13T17:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.647027 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.647127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.647145 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.647172 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.647188 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:57Z","lastTransitionTime":"2025-12-13T17:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.749664 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.749734 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.749751 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.749776 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.749821 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:57Z","lastTransitionTime":"2025-12-13T17:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.852543 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.852617 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.852642 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.852674 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.852696 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:57Z","lastTransitionTime":"2025-12-13T17:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.955195 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.955255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.955277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.955304 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:57 crc kubenswrapper[4989]: I1213 17:17:57.955326 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:57Z","lastTransitionTime":"2025-12-13T17:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.013813 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.013840 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:17:58 crc kubenswrapper[4989]: E1213 17:17:58.013948 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:17:58 crc kubenswrapper[4989]: E1213 17:17:58.014078 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.013829 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:17:58 crc kubenswrapper[4989]: E1213 17:17:58.014152 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.057332 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.057397 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.057415 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.057439 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.057461 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.161028 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.161099 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.161120 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.161148 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.161169 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.264122 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.264160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.264176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.264198 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.264215 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.366434 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.366493 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.366509 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.366536 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.366554 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.468960 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.469021 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.469041 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.469067 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.469088 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.572319 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.572404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.572436 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.572475 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.572499 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.675991 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.676070 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.676098 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.676129 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.676147 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.778748 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.778814 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.778830 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.778848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.778860 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.846250 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.846298 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.846315 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.846342 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.846364 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: E1213 17:17:58.868484 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.872921 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.872976 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.872989 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.873005 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.873016 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: E1213 17:17:58.888727 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.897869 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.897956 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.897977 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.898011 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.898032 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: E1213 17:17:58.915013 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.919434 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.919472 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.919482 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.919500 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.919509 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: E1213 17:17:58.929345 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.932025 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.932072 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.932087 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.932110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.932126 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:58 crc kubenswrapper[4989]: E1213 17:17:58.950075 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"9e98691d-d678-4159-b20c-43488bb99dd0\\\",\\\"systemUUID\\\":\\\"364e0ead-9bda-403b-8614-341ecd5845fc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:17:58Z is after 2025-08-24T17:21:41Z" Dec 13 17:17:58 crc kubenswrapper[4989]: E1213 17:17:58.950259 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.952208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.952260 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.952277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.952302 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:58 crc kubenswrapper[4989]: I1213 17:17:58.952319 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:58Z","lastTransitionTime":"2025-12-13T17:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.013838 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:17:59 crc kubenswrapper[4989]: E1213 17:17:59.014045 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.055286 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.055350 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.055368 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.055393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.055411 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:59Z","lastTransitionTime":"2025-12-13T17:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.157659 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.157699 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.157710 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.157729 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.157740 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:59Z","lastTransitionTime":"2025-12-13T17:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.260021 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.260056 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.260067 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.260082 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.260093 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:59Z","lastTransitionTime":"2025-12-13T17:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.362676 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.362716 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.362728 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.362744 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.362753 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:59Z","lastTransitionTime":"2025-12-13T17:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.464668 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.464748 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.464771 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.464832 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.464868 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:59Z","lastTransitionTime":"2025-12-13T17:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.566978 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.567031 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.567048 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.567070 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.567087 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:59Z","lastTransitionTime":"2025-12-13T17:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.669849 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.669905 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.669925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.669949 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.669965 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:59Z","lastTransitionTime":"2025-12-13T17:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.772729 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.772874 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.772897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.772926 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.772947 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:59Z","lastTransitionTime":"2025-12-13T17:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.876898 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.876975 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.876998 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.877032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.877060 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:59Z","lastTransitionTime":"2025-12-13T17:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.980985 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.981060 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.981078 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.981105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:17:59 crc kubenswrapper[4989]: I1213 17:17:59.981125 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:17:59Z","lastTransitionTime":"2025-12-13T17:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.013699 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.013860 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.013936 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:00 crc kubenswrapper[4989]: E1213 17:18:00.014041 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:00 crc kubenswrapper[4989]: E1213 17:18:00.014159 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:00 crc kubenswrapper[4989]: E1213 17:18:00.014732 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.084679 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.084764 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.084831 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.084929 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.084958 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:00Z","lastTransitionTime":"2025-12-13T17:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.189268 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.189311 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.189322 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.189338 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.189352 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:00Z","lastTransitionTime":"2025-12-13T17:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.292034 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.292435 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.292518 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.292598 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.292643 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:00Z","lastTransitionTime":"2025-12-13T17:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.396127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.396191 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.396201 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.396215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.396225 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:00Z","lastTransitionTime":"2025-12-13T17:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.498447 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.498487 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.498500 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.498517 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.498530 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:00Z","lastTransitionTime":"2025-12-13T17:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.601072 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.601104 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.601112 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.601123 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.601132 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:00Z","lastTransitionTime":"2025-12-13T17:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.703673 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.703739 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.703756 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.703781 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.703837 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:00Z","lastTransitionTime":"2025-12-13T17:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.806049 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.806105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.806123 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.806149 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.806167 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:00Z","lastTransitionTime":"2025-12-13T17:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.910097 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.910146 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.910157 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.910175 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:00 crc kubenswrapper[4989]: I1213 17:18:00.910187 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:00Z","lastTransitionTime":"2025-12-13T17:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.013122 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.013195 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.013214 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.013246 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.013265 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:01Z","lastTransitionTime":"2025-12-13T17:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.013543 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:01 crc kubenswrapper[4989]: E1213 17:18:01.013922 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.116373 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.116433 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.116450 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.116476 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.116492 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:01Z","lastTransitionTime":"2025-12-13T17:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.219912 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.220027 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.220091 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.220110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.220123 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:01Z","lastTransitionTime":"2025-12-13T17:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.322874 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.322931 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.322946 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.322966 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.322980 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:01Z","lastTransitionTime":"2025-12-13T17:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.425674 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.425727 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.425740 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.425760 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.425777 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:01Z","lastTransitionTime":"2025-12-13T17:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.528938 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.528994 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.529020 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.529047 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.529063 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:01Z","lastTransitionTime":"2025-12-13T17:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.632603 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.632642 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.632658 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.632678 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.632691 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:01Z","lastTransitionTime":"2025-12-13T17:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.735597 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.735659 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.735678 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.735701 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.735718 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:01Z","lastTransitionTime":"2025-12-13T17:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.837449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.837509 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.837526 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.837549 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.837565 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:01Z","lastTransitionTime":"2025-12-13T17:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.939717 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.939819 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.939848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.939874 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:01 crc kubenswrapper[4989]: I1213 17:18:01.939892 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:01Z","lastTransitionTime":"2025-12-13T17:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.013842 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.013958 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.014032 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:02 crc kubenswrapper[4989]: E1213 17:18:02.014175 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:02 crc kubenswrapper[4989]: E1213 17:18:02.014331 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:02 crc kubenswrapper[4989]: E1213 17:18:02.014435 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.042558 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.042602 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.042611 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.042627 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.042636 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:02Z","lastTransitionTime":"2025-12-13T17:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.145279 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.145425 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.145455 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.145499 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.145526 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:02Z","lastTransitionTime":"2025-12-13T17:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.248019 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.248084 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.248098 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.248113 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.248123 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:02Z","lastTransitionTime":"2025-12-13T17:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.350308 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.350394 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.350421 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.350463 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.350489 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:02Z","lastTransitionTime":"2025-12-13T17:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.453442 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.453497 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.453506 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.453524 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.453534 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:02Z","lastTransitionTime":"2025-12-13T17:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.555589 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.555641 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.555654 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.555671 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.555680 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:02Z","lastTransitionTime":"2025-12-13T17:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.658665 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.658729 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.658752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.658782 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.658850 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:02Z","lastTransitionTime":"2025-12-13T17:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.762373 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.762460 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.762487 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.762518 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.762538 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:02Z","lastTransitionTime":"2025-12-13T17:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.866467 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.866541 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.866567 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.866599 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.866622 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:02Z","lastTransitionTime":"2025-12-13T17:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.969668 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.969717 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.969729 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.969752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:02 crc kubenswrapper[4989]: I1213 17:18:02.969762 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:02Z","lastTransitionTime":"2025-12-13T17:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.014451 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:03 crc kubenswrapper[4989]: E1213 17:18:03.014771 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.073586 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.073637 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.073650 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.073673 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.073688 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:03Z","lastTransitionTime":"2025-12-13T17:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.177970 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.178037 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.178051 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.178073 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.178091 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:03Z","lastTransitionTime":"2025-12-13T17:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.281536 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.281611 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.281632 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.281664 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.281692 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:03Z","lastTransitionTime":"2025-12-13T17:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.390274 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.390383 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.390402 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.390430 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.390450 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:03Z","lastTransitionTime":"2025-12-13T17:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.493158 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.493232 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.493252 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.493283 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.493304 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:03Z","lastTransitionTime":"2025-12-13T17:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.596433 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.596512 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.596536 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.596564 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.596581 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:03Z","lastTransitionTime":"2025-12-13T17:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.699316 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.699413 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.699438 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.699476 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.699503 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:03Z","lastTransitionTime":"2025-12-13T17:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.803234 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.803311 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.803337 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.803365 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.803383 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:03Z","lastTransitionTime":"2025-12-13T17:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.906694 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.906740 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.906751 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.906814 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:03 crc kubenswrapper[4989]: I1213 17:18:03.906827 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:03Z","lastTransitionTime":"2025-12-13T17:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.010152 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.010280 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.010307 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.010343 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.010369 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:04Z","lastTransitionTime":"2025-12-13T17:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.014064 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.014121 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:04 crc kubenswrapper[4989]: E1213 17:18:04.014328 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.014453 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:04 crc kubenswrapper[4989]: E1213 17:18:04.014494 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:04 crc kubenswrapper[4989]: E1213 17:18:04.014707 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.113345 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.113407 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.113425 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.113450 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.113468 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:04Z","lastTransitionTime":"2025-12-13T17:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.216648 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.216689 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.216700 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.216715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.216724 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:04Z","lastTransitionTime":"2025-12-13T17:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.320273 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.320316 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.320326 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.320342 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.320352 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:04Z","lastTransitionTime":"2025-12-13T17:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.424032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.424135 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.424153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.424186 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.424208 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:04Z","lastTransitionTime":"2025-12-13T17:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.527450 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.527502 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.527515 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.527532 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.527544 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:04Z","lastTransitionTime":"2025-12-13T17:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.630185 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.630231 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.630243 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.630261 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.630274 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:04Z","lastTransitionTime":"2025-12-13T17:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.732708 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.732759 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.732769 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.732781 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.732804 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:04Z","lastTransitionTime":"2025-12-13T17:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.836327 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.836404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.836431 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.836462 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.836488 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:04Z","lastTransitionTime":"2025-12-13T17:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.939282 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.939333 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.939344 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.939362 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:04 crc kubenswrapper[4989]: I1213 17:18:04.939374 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:04Z","lastTransitionTime":"2025-12-13T17:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.013874 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:05 crc kubenswrapper[4989]: E1213 17:18:05.014089 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.042980 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.043038 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.043056 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.043086 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.043103 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:05Z","lastTransitionTime":"2025-12-13T17:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.146471 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.146522 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.146539 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.146560 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.146578 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:05Z","lastTransitionTime":"2025-12-13T17:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.249639 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.249687 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.249701 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.249718 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.249731 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:05Z","lastTransitionTime":"2025-12-13T17:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.351853 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.351898 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.351914 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.351935 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.351949 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:05Z","lastTransitionTime":"2025-12-13T17:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.455292 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.455334 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.455351 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.455376 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.455395 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:05Z","lastTransitionTime":"2025-12-13T17:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.559324 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.559415 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.559442 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.559474 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.559496 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:05Z","lastTransitionTime":"2025-12-13T17:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.663080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.663143 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.663161 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.663188 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.663209 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:05Z","lastTransitionTime":"2025-12-13T17:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.765843 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.765924 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.765941 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.765991 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.766006 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:05Z","lastTransitionTime":"2025-12-13T17:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.869312 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.869351 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.869360 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.869375 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.869384 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:05Z","lastTransitionTime":"2025-12-13T17:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.973460 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.973527 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.973541 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.973561 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:05 crc kubenswrapper[4989]: I1213 17:18:05.973574 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:05Z","lastTransitionTime":"2025-12-13T17:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.014370 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.014438 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.014405 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:06 crc kubenswrapper[4989]: E1213 17:18:06.014552 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:06 crc kubenswrapper[4989]: E1213 17:18:06.014836 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:06 crc kubenswrapper[4989]: E1213 17:18:06.014692 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.035482 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.054117 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a29bcd2fa11e91cbe04b08cddabe38ece02d538851a71eb618b39472508793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.073231 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2b01148-171a-4f86-84a7-d326739e0dcf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e83ac472539205ec902034bdd5d9845e14cb1d03ca65088854fcfb76902abd5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dn7p6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-nh9k2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.076736 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.076782 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.076814 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.076870 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.076883 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:06Z","lastTransitionTime":"2025-12-13T17:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.099400 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"101724b9-153f-4f9d-849a-c04a343e7446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:41Z\\\",\\\"message\\\":\\\"Family:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-config-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.161\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1213 17:17:41.791836 6994 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nI1213 17:17:41.791993 6994 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:17:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blhj7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-z7249\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.113825 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d912915-788e-412a-bae8-6eccd6b4c238\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45wzg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:05Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lfpf8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.144708 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f90a95c3-4a07-4fa4-99cf-14fe2f935490\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3467bc40c0c255a3531365541bb1b016355ff39d186c2584ed9da8213fe8a0bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da3288347bc3ea9905194fcdf723a5ce766cd0d5d29dca4ae07723e6b79f56ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1382d2c1b354c41acd3018549975737e25a51c14f8fcfa775982f77ddc351d2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://825d9a66e1908d34257c9353caac6c3e69b926da3902dfea5e06a2b51c66c516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05ba18c02cfa1a75610fbe52f3d0d5c3f71f13a5917f5a7421083c52c08f0379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f13e8226792e2a5ea902ca35088062e451f123ea755913e516fb5cf4df5a008\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83e63519cb50c5a36cae8cfdac7343717f6aef2d52dd96ff9dde9a665024014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f3bfdf0a61b1de221f6ae6ad24b668c206f160c9fe3d2a8d4ebe098e7e0ffb9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.159175 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.172503 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30798eda-9ae1-43fd-918b-a13c4be89568\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e29c7ec9865bdfb6171658395dfbc625887b86c8a5bb13934d599e428e4d9827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73d5adf20751d8f4b42d8626fde2395ff599df5899e080ee9bdda5022f423ef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://060180d6fdbf003d76131d13c23721fe8c66388851bca8bccc9d6807246c8f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1292311ae4317f9787d367aba8c45e2d8bcb3139f7950aa6ec42e379782f325d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.180554 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.180601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.180614 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.180637 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.180654 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:06Z","lastTransitionTime":"2025-12-13T17:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.187513 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8d32e3a5112b363d58014eca02d567e51ead3cdd0395229ecbc0c3b7cbef2b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://442afde82e899861439ffc492b34e28a20b4d1a043971ed83a60b91ed44f9d31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.204087 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-hllvq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-13T17:17:37Z\\\",\\\"message\\\":\\\"2025-12-13T17:16:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8\\\\n2025-12-13T17:16:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_127a022a-76f2-4ba6-a274-cff8a64b41c8 to /host/opt/cni/bin/\\\\n2025-12-13T17:16:52Z [verbose] multus-daemon started\\\\n2025-12-13T17:16:52Z [verbose] Readiness Indicator file check\\\\n2025-12-13T17:17:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6hpvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hllvq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.218052 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vmx98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25dd508e-594b-4f32-af84-61ee8d65f38b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec5dc526076818863fb9b301e3ba928c64c2df5ff333b742d270a01c74d9491\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24j5k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:55Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vmx98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.232818 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf576839-f84e-436f-8855-d0027a0c6ee4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:17:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6692a174856e912e26f21402b25c692b8b4c91da2f4433ec3a18a818a2efdfa8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99a051a4354d997b049d0168aceafd845476444acd18b75d3af560cbf459ff6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:17:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwxcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:17:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tbb8h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.244433 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7c10df0-6a5f-4a82-9628-d2d77c6c070a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a3723ba13bbec66d7eeec936820edf9373e18d0d7d29ab050945b45106f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a86b79af750721eefcf69193011245de1f2146f9022c876ddc4f64644de04a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b37caaf35d75e9c78c3cfb36cf9be761819cef7152502470ecd133f8c4d3edc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.256042 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d12847a-6be0-4ab1-8052-cd417f1525c9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-13T17:16:38Z\\\",\\\"message\\\":\\\"W1213 17:16:28.183879 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1213 17:16:28.184166 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765646188 cert, and key in /tmp/serving-cert-2596394001/serving-signer.crt, /tmp/serving-cert-2596394001/serving-signer.key\\\\nI1213 17:16:28.416608 1 observer_polling.go:159] Starting file observer\\\\nW1213 17:16:28.420009 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1213 17:16:28.420157 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1213 17:16:28.420818 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2596394001/tls.crt::/tmp/serving-cert-2596394001/tls.key\\\\\\\"\\\\nF1213 17:16:38.669257 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.269086 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85df4a7dee922b10aa59199a29a3a9c26c46414b13953af3b3da4a14b75ed529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.283578 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.283627 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.283638 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.283658 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.283671 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:06Z","lastTransitionTime":"2025-12-13T17:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.288764 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.309207 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ed6e54-65fe-4383-9578-d0c89a69ecec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73ec1928c51faec5e708977671b9001367591c5f03cf58bccae468c8e116cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0a44b7112100b7f8f908ec69f93e80cb8d25153378ecc850fce3b66c43894fb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec980d1accd272c81fda30c408f94fe89fd6f57106c16e4007c205cf77a15565\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e2609aed911517bfa311833418fce3e431cd90fb0ed2a82893bb271239b965f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac6c877a973f63993b9a7de99b9fcdedb2b8e57f48b451c7cb1357e3502825eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24d3096a471731d7552f93727a59724ad2c8e6cc9e8424eea31ae8a9a6ea912a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://67b6847edfe1a515c47e808becfc15d94a87c447dac279a1880516eafb2d9410\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdsjc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-dv9sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.327068 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4l98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"851885b5-3fdc-4e01-87d2-4a79a73acd6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f3a166d307bedd24e1ba2a4d9137b510c6fb8fbfac830680fbb12a2a6d84a0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vcv48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:51Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4l98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.350677 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"390325ff-bbf3-4c16-88c7-cdb04976409a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961b7ddbae6e098a3f15adc1253a50fc94fcd98c362f248e976570a9a391769f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-13T17:16:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://386f0a89a932893d541d90c3daec49677e1ae574c5bc75138187d7f67b0b0c9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://386f0a89a932893d541d90c3daec49677e1ae574c5bc75138187d7f67b0b0c9a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-13T17:16:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-13T17:16:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-13T17:16:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-13T17:18:06Z is after 2025-08-24T17:21:41Z" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.386325 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.386360 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.386369 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.386382 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.386393 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:06Z","lastTransitionTime":"2025-12-13T17:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.488312 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.488700 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.488713 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.488729 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.488742 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:06Z","lastTransitionTime":"2025-12-13T17:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.591489 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.591544 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.591561 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.591584 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.591602 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:06Z","lastTransitionTime":"2025-12-13T17:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.694380 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.694436 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.694458 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.694489 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.694513 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:06Z","lastTransitionTime":"2025-12-13T17:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.797286 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.797329 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.797343 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.797359 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.797371 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:06Z","lastTransitionTime":"2025-12-13T17:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.899780 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.899833 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.899841 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.899853 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:06 crc kubenswrapper[4989]: I1213 17:18:06.899863 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:06Z","lastTransitionTime":"2025-12-13T17:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.002771 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.002857 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.002900 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.002920 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.002937 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:07Z","lastTransitionTime":"2025-12-13T17:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.014047 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:07 crc kubenswrapper[4989]: E1213 17:18:07.014177 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.106144 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.106218 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.106243 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.106277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.106302 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:07Z","lastTransitionTime":"2025-12-13T17:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.209498 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.209553 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.209569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.209589 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.209607 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:07Z","lastTransitionTime":"2025-12-13T17:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.311518 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.311588 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.311605 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.311631 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.311652 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:07Z","lastTransitionTime":"2025-12-13T17:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.413521 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.413581 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.413601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.413626 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.413643 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:07Z","lastTransitionTime":"2025-12-13T17:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.516842 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.516891 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.516909 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.516931 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.516947 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:07Z","lastTransitionTime":"2025-12-13T17:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.619593 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.619642 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.619651 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.619666 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.619676 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:07Z","lastTransitionTime":"2025-12-13T17:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.722558 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.722599 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.722608 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.722623 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.722633 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:07Z","lastTransitionTime":"2025-12-13T17:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.825203 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.825253 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.825270 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.825293 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.825309 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:07Z","lastTransitionTime":"2025-12-13T17:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.928383 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.928464 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.928485 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.928515 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:07 crc kubenswrapper[4989]: I1213 17:18:07.928537 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:07Z","lastTransitionTime":"2025-12-13T17:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.013889 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.013946 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.013998 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:08 crc kubenswrapper[4989]: E1213 17:18:08.014082 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:08 crc kubenswrapper[4989]: E1213 17:18:08.014260 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:08 crc kubenswrapper[4989]: E1213 17:18:08.014466 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.031258 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.031317 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.031343 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.031371 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.031393 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:08Z","lastTransitionTime":"2025-12-13T17:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.134368 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.134439 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.134456 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.134478 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.134494 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:08Z","lastTransitionTime":"2025-12-13T17:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.238206 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.238271 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.238289 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.238312 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.238329 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:08Z","lastTransitionTime":"2025-12-13T17:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.341939 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.342028 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.342122 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.342148 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.342167 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:08Z","lastTransitionTime":"2025-12-13T17:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.445698 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.445783 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.445883 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.445915 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.445934 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:08Z","lastTransitionTime":"2025-12-13T17:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.548456 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.548584 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.548608 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.548638 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.548661 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:08Z","lastTransitionTime":"2025-12-13T17:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.650751 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.650817 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.650834 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.650851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.650863 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:08Z","lastTransitionTime":"2025-12-13T17:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.753042 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.753094 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.753108 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.753125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.753139 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:08Z","lastTransitionTime":"2025-12-13T17:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.856199 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.856255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.856272 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.856293 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.856310 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:08Z","lastTransitionTime":"2025-12-13T17:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.959027 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.959107 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.959142 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.959172 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:08 crc kubenswrapper[4989]: I1213 17:18:08.959191 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:08Z","lastTransitionTime":"2025-12-13T17:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.014268 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:09 crc kubenswrapper[4989]: E1213 17:18:09.014465 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.061518 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.061617 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.061641 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.061671 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.061693 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:09Z","lastTransitionTime":"2025-12-13T17:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.155274 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.155348 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.155364 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.155388 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.155405 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-13T17:18:09Z","lastTransitionTime":"2025-12-13T17:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.213995 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x"] Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.214327 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.216685 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.216947 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.217072 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.217098 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.263285 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tbb8h" podStartSLOduration=78.263219127 podStartE2EDuration="1m18.263219127s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:09.247445797 +0000 UTC m=+103.853892945" watchObservedRunningTime="2025-12-13 17:18:09.263219127 +0000 UTC m=+103.869666285" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.294025 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=86.294007674 podStartE2EDuration="1m26.294007674s" podCreationTimestamp="2025-12-13 17:16:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:09.277836253 +0000 UTC m=+103.884283401" watchObservedRunningTime="2025-12-13 17:18:09.294007674 +0000 UTC m=+103.900454812" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.307631 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=49.30760816 podStartE2EDuration="49.30760816s" podCreationTimestamp="2025-12-13 17:17:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:09.294106147 +0000 UTC m=+103.900553295" watchObservedRunningTime="2025-12-13 17:18:09.30760816 +0000 UTC m=+103.914055298" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.337069 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-hllvq" podStartSLOduration=78.337048878 podStartE2EDuration="1m18.337048878s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:09.322499704 +0000 UTC m=+103.928946842" watchObservedRunningTime="2025-12-13 17:18:09.337048878 +0000 UTC m=+103.943496036" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.347639 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-vmx98" podStartSLOduration=79.347623173 podStartE2EDuration="1m19.347623173s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:09.337336117 +0000 UTC m=+103.943783255" watchObservedRunningTime="2025-12-13 17:18:09.347623173 +0000 UTC m=+103.954070311" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.359524 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.359500988 podStartE2EDuration="25.359500988s" podCreationTimestamp="2025-12-13 17:17:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:09.359452116 +0000 UTC m=+103.965899254" watchObservedRunningTime="2025-12-13 17:18:09.359500988 +0000 UTC m=+103.965948126" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.359978 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-n4l98" podStartSLOduration=79.359971712 podStartE2EDuration="1m19.359971712s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:09.348114368 +0000 UTC m=+103.954561506" watchObservedRunningTime="2025-12-13 17:18:09.359971712 +0000 UTC m=+103.966418850" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.368679 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/551bbe25-16a9-4547-9c9c-defa6e62966a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.368751 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/551bbe25-16a9-4547-9c9c-defa6e62966a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.368777 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/551bbe25-16a9-4547-9c9c-defa6e62966a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.368853 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/551bbe25-16a9-4547-9c9c-defa6e62966a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.368893 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.368912 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/551bbe25-16a9-4547-9c9c-defa6e62966a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: E1213 17:18:09.369031 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:18:09 crc kubenswrapper[4989]: E1213 17:18:09.369088 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs podName:7d912915-788e-412a-bae8-6eccd6b4c238 nodeName:}" failed. No retries permitted until 2025-12-13 17:19:13.369074292 +0000 UTC m=+167.975521430 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs") pod "network-metrics-daemon-lfpf8" (UID: "7d912915-788e-412a-bae8-6eccd6b4c238") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.390653 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=85.390636796 podStartE2EDuration="1m25.390636796s" podCreationTimestamp="2025-12-13 17:16:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:09.375919497 +0000 UTC m=+103.982366645" watchObservedRunningTime="2025-12-13 17:18:09.390636796 +0000 UTC m=+103.997083934" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.418386 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-dv9sz" podStartSLOduration=78.418349052 podStartE2EDuration="1m18.418349052s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:09.417999582 +0000 UTC m=+104.024446720" watchObservedRunningTime="2025-12-13 17:18:09.418349052 +0000 UTC m=+104.024796190" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.451408 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=81.451382187 podStartE2EDuration="1m21.451382187s" podCreationTimestamp="2025-12-13 17:16:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:09.451185401 +0000 UTC m=+104.057632549" watchObservedRunningTime="2025-12-13 17:18:09.451382187 +0000 UTC m=+104.057829325" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.470329 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/551bbe25-16a9-4547-9c9c-defa6e62966a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.470398 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/551bbe25-16a9-4547-9c9c-defa6e62966a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.470426 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/551bbe25-16a9-4547-9c9c-defa6e62966a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.470454 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/551bbe25-16a9-4547-9c9c-defa6e62966a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.470515 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/551bbe25-16a9-4547-9c9c-defa6e62966a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.471469 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/551bbe25-16a9-4547-9c9c-defa6e62966a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.471594 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/551bbe25-16a9-4547-9c9c-defa6e62966a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.472607 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/551bbe25-16a9-4547-9c9c-defa6e62966a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.476512 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/551bbe25-16a9-4547-9c9c-defa6e62966a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.488545 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/551bbe25-16a9-4547-9c9c-defa6e62966a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-q6k7x\" (UID: \"551bbe25-16a9-4547-9c9c-defa6e62966a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.530746 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podStartSLOduration=78.530727383 podStartE2EDuration="1m18.530727383s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:09.499325986 +0000 UTC m=+104.105773124" watchObservedRunningTime="2025-12-13 17:18:09.530727383 +0000 UTC m=+104.137174521" Dec 13 17:18:09 crc kubenswrapper[4989]: I1213 17:18:09.531116 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" Dec 13 17:18:10 crc kubenswrapper[4989]: I1213 17:18:10.013986 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:10 crc kubenswrapper[4989]: I1213 17:18:10.013994 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:10 crc kubenswrapper[4989]: I1213 17:18:10.014042 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:10 crc kubenswrapper[4989]: E1213 17:18:10.014468 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:10 crc kubenswrapper[4989]: E1213 17:18:10.014690 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:10 crc kubenswrapper[4989]: I1213 17:18:10.014770 4989 scope.go:117] "RemoveContainer" containerID="8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1" Dec 13 17:18:10 crc kubenswrapper[4989]: E1213 17:18:10.014810 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:10 crc kubenswrapper[4989]: E1213 17:18:10.014955 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-z7249_openshift-ovn-kubernetes(101724b9-153f-4f9d-849a-c04a343e7446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" Dec 13 17:18:10 crc kubenswrapper[4989]: I1213 17:18:10.534006 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" event={"ID":"551bbe25-16a9-4547-9c9c-defa6e62966a","Type":"ContainerStarted","Data":"6637e8a7a84c46448347c104b60989ce5027eb2e0228b40fb29721c8e590599b"} Dec 13 17:18:10 crc kubenswrapper[4989]: I1213 17:18:10.534050 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" event={"ID":"551bbe25-16a9-4547-9c9c-defa6e62966a","Type":"ContainerStarted","Data":"cd6dac1186adc7e535ecaf3230e2def68db0fde155d96d940d5e0faaf5a532ec"} Dec 13 17:18:10 crc kubenswrapper[4989]: I1213 17:18:10.547660 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-q6k7x" podStartSLOduration=80.547646632 podStartE2EDuration="1m20.547646632s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:10.545220939 +0000 UTC m=+105.151668077" watchObservedRunningTime="2025-12-13 17:18:10.547646632 +0000 UTC m=+105.154093770" Dec 13 17:18:11 crc kubenswrapper[4989]: I1213 17:18:11.014018 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:11 crc kubenswrapper[4989]: E1213 17:18:11.014130 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:12 crc kubenswrapper[4989]: I1213 17:18:12.014079 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:12 crc kubenswrapper[4989]: I1213 17:18:12.014152 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:12 crc kubenswrapper[4989]: I1213 17:18:12.014177 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:12 crc kubenswrapper[4989]: E1213 17:18:12.015151 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:12 crc kubenswrapper[4989]: E1213 17:18:12.015259 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:12 crc kubenswrapper[4989]: E1213 17:18:12.015305 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:13 crc kubenswrapper[4989]: I1213 17:18:13.014113 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:13 crc kubenswrapper[4989]: E1213 17:18:13.014578 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:14 crc kubenswrapper[4989]: I1213 17:18:14.014238 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:14 crc kubenswrapper[4989]: I1213 17:18:14.014323 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:14 crc kubenswrapper[4989]: I1213 17:18:14.014354 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:14 crc kubenswrapper[4989]: E1213 17:18:14.014403 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:14 crc kubenswrapper[4989]: E1213 17:18:14.014501 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:14 crc kubenswrapper[4989]: E1213 17:18:14.014612 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:15 crc kubenswrapper[4989]: I1213 17:18:15.013598 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:15 crc kubenswrapper[4989]: E1213 17:18:15.013859 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:16 crc kubenswrapper[4989]: I1213 17:18:16.013929 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:16 crc kubenswrapper[4989]: I1213 17:18:16.013981 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:16 crc kubenswrapper[4989]: I1213 17:18:16.013953 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:16 crc kubenswrapper[4989]: E1213 17:18:16.015155 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:16 crc kubenswrapper[4989]: E1213 17:18:16.015556 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:16 crc kubenswrapper[4989]: E1213 17:18:16.015652 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:17 crc kubenswrapper[4989]: I1213 17:18:17.014101 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:17 crc kubenswrapper[4989]: E1213 17:18:17.014300 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:18 crc kubenswrapper[4989]: I1213 17:18:18.014188 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:18 crc kubenswrapper[4989]: I1213 17:18:18.014256 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:18 crc kubenswrapper[4989]: I1213 17:18:18.014216 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:18 crc kubenswrapper[4989]: E1213 17:18:18.014447 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:18 crc kubenswrapper[4989]: E1213 17:18:18.014763 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:18 crc kubenswrapper[4989]: E1213 17:18:18.014852 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:19 crc kubenswrapper[4989]: I1213 17:18:19.014147 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:19 crc kubenswrapper[4989]: E1213 17:18:19.014447 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:20 crc kubenswrapper[4989]: I1213 17:18:20.013595 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:20 crc kubenswrapper[4989]: I1213 17:18:20.013635 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:20 crc kubenswrapper[4989]: E1213 17:18:20.013780 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:20 crc kubenswrapper[4989]: I1213 17:18:20.014014 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:20 crc kubenswrapper[4989]: E1213 17:18:20.014116 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:20 crc kubenswrapper[4989]: E1213 17:18:20.014343 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:21 crc kubenswrapper[4989]: I1213 17:18:21.013462 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:21 crc kubenswrapper[4989]: E1213 17:18:21.013688 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:22 crc kubenswrapper[4989]: I1213 17:18:22.014338 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:22 crc kubenswrapper[4989]: E1213 17:18:22.014462 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:22 crc kubenswrapper[4989]: I1213 17:18:22.014543 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:22 crc kubenswrapper[4989]: I1213 17:18:22.014351 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:22 crc kubenswrapper[4989]: E1213 17:18:22.014699 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:22 crc kubenswrapper[4989]: E1213 17:18:22.015089 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:23 crc kubenswrapper[4989]: I1213 17:18:23.014090 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:23 crc kubenswrapper[4989]: E1213 17:18:23.014287 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:23 crc kubenswrapper[4989]: I1213 17:18:23.015527 4989 scope.go:117] "RemoveContainer" containerID="8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1" Dec 13 17:18:23 crc kubenswrapper[4989]: I1213 17:18:23.579450 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/3.log" Dec 13 17:18:23 crc kubenswrapper[4989]: I1213 17:18:23.582016 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerStarted","Data":"14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa"} Dec 13 17:18:23 crc kubenswrapper[4989]: I1213 17:18:23.582356 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:18:23 crc kubenswrapper[4989]: I1213 17:18:23.609247 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podStartSLOduration=92.609231847 podStartE2EDuration="1m32.609231847s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:23.608081672 +0000 UTC m=+118.214528820" watchObservedRunningTime="2025-12-13 17:18:23.609231847 +0000 UTC m=+118.215678985" Dec 13 17:18:23 crc kubenswrapper[4989]: I1213 17:18:23.871715 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lfpf8"] Dec 13 17:18:23 crc kubenswrapper[4989]: I1213 17:18:23.871865 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:23 crc kubenswrapper[4989]: E1213 17:18:23.871973 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:24 crc kubenswrapper[4989]: I1213 17:18:24.013768 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:24 crc kubenswrapper[4989]: I1213 17:18:24.013842 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:24 crc kubenswrapper[4989]: E1213 17:18:24.014354 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:24 crc kubenswrapper[4989]: I1213 17:18:24.013890 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:24 crc kubenswrapper[4989]: E1213 17:18:24.014534 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:24 crc kubenswrapper[4989]: E1213 17:18:24.014579 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:24 crc kubenswrapper[4989]: I1213 17:18:24.586179 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hllvq_1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf/kube-multus/1.log" Dec 13 17:18:24 crc kubenswrapper[4989]: I1213 17:18:24.586610 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hllvq_1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf/kube-multus/0.log" Dec 13 17:18:24 crc kubenswrapper[4989]: I1213 17:18:24.586655 4989 generic.go:334] "Generic (PLEG): container finished" podID="1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf" containerID="4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c" exitCode=1 Dec 13 17:18:24 crc kubenswrapper[4989]: I1213 17:18:24.586728 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hllvq" event={"ID":"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf","Type":"ContainerDied","Data":"4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c"} Dec 13 17:18:24 crc kubenswrapper[4989]: I1213 17:18:24.586784 4989 scope.go:117] "RemoveContainer" containerID="d350bfa18c242eab74da5b1a54719217b55226b30c253b999d95ee02dbee9494" Dec 13 17:18:24 crc kubenswrapper[4989]: I1213 17:18:24.587320 4989 scope.go:117] "RemoveContainer" containerID="4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c" Dec 13 17:18:24 crc kubenswrapper[4989]: E1213 17:18:24.587493 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-hllvq_openshift-multus(1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf)\"" pod="openshift-multus/multus-hllvq" podUID="1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf" Dec 13 17:18:25 crc kubenswrapper[4989]: I1213 17:18:25.592602 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hllvq_1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf/kube-multus/1.log" Dec 13 17:18:25 crc kubenswrapper[4989]: E1213 17:18:25.980303 4989 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 13 17:18:26 crc kubenswrapper[4989]: I1213 17:18:26.014559 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:26 crc kubenswrapper[4989]: I1213 17:18:26.014617 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:26 crc kubenswrapper[4989]: I1213 17:18:26.014559 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:26 crc kubenswrapper[4989]: I1213 17:18:26.014661 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:26 crc kubenswrapper[4989]: E1213 17:18:26.015902 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:26 crc kubenswrapper[4989]: E1213 17:18:26.016014 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:26 crc kubenswrapper[4989]: E1213 17:18:26.017000 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:26 crc kubenswrapper[4989]: E1213 17:18:26.017194 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:26 crc kubenswrapper[4989]: E1213 17:18:26.108975 4989 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 13 17:18:28 crc kubenswrapper[4989]: I1213 17:18:28.014415 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:28 crc kubenswrapper[4989]: I1213 17:18:28.014462 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:28 crc kubenswrapper[4989]: I1213 17:18:28.014528 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:28 crc kubenswrapper[4989]: E1213 17:18:28.014628 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:28 crc kubenswrapper[4989]: I1213 17:18:28.014654 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:28 crc kubenswrapper[4989]: E1213 17:18:28.014783 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:28 crc kubenswrapper[4989]: E1213 17:18:28.014985 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:28 crc kubenswrapper[4989]: E1213 17:18:28.015067 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:30 crc kubenswrapper[4989]: I1213 17:18:30.014600 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:30 crc kubenswrapper[4989]: I1213 17:18:30.014658 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:30 crc kubenswrapper[4989]: E1213 17:18:30.014706 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:30 crc kubenswrapper[4989]: I1213 17:18:30.014600 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:30 crc kubenswrapper[4989]: E1213 17:18:30.014889 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:30 crc kubenswrapper[4989]: E1213 17:18:30.014942 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:30 crc kubenswrapper[4989]: I1213 17:18:30.015064 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:30 crc kubenswrapper[4989]: E1213 17:18:30.015108 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:31 crc kubenswrapper[4989]: E1213 17:18:31.111216 4989 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 13 17:18:32 crc kubenswrapper[4989]: I1213 17:18:32.014365 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:32 crc kubenswrapper[4989]: I1213 17:18:32.014477 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:32 crc kubenswrapper[4989]: E1213 17:18:32.014510 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:32 crc kubenswrapper[4989]: I1213 17:18:32.014373 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:32 crc kubenswrapper[4989]: I1213 17:18:32.014375 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:32 crc kubenswrapper[4989]: E1213 17:18:32.014653 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:32 crc kubenswrapper[4989]: E1213 17:18:32.014746 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:32 crc kubenswrapper[4989]: E1213 17:18:32.014915 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:34 crc kubenswrapper[4989]: I1213 17:18:34.014479 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:34 crc kubenswrapper[4989]: E1213 17:18:34.015108 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:34 crc kubenswrapper[4989]: I1213 17:18:34.014520 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:34 crc kubenswrapper[4989]: I1213 17:18:34.014715 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:34 crc kubenswrapper[4989]: I1213 17:18:34.014534 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:34 crc kubenswrapper[4989]: E1213 17:18:34.015423 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:34 crc kubenswrapper[4989]: E1213 17:18:34.015512 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:34 crc kubenswrapper[4989]: E1213 17:18:34.015552 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:36 crc kubenswrapper[4989]: I1213 17:18:36.014039 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:36 crc kubenswrapper[4989]: I1213 17:18:36.014039 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:36 crc kubenswrapper[4989]: E1213 17:18:36.016210 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:36 crc kubenswrapper[4989]: I1213 17:18:36.016353 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:36 crc kubenswrapper[4989]: E1213 17:18:36.016528 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:36 crc kubenswrapper[4989]: I1213 17:18:36.016680 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:36 crc kubenswrapper[4989]: E1213 17:18:36.016967 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:36 crc kubenswrapper[4989]: E1213 17:18:36.017183 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:36 crc kubenswrapper[4989]: E1213 17:18:36.112302 4989 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 13 17:18:38 crc kubenswrapper[4989]: I1213 17:18:38.014522 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:38 crc kubenswrapper[4989]: I1213 17:18:38.014626 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:38 crc kubenswrapper[4989]: I1213 17:18:38.014550 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:38 crc kubenswrapper[4989]: E1213 17:18:38.014665 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:38 crc kubenswrapper[4989]: I1213 17:18:38.014784 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:38 crc kubenswrapper[4989]: E1213 17:18:38.014817 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:38 crc kubenswrapper[4989]: E1213 17:18:38.015089 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:38 crc kubenswrapper[4989]: E1213 17:18:38.015181 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:39 crc kubenswrapper[4989]: I1213 17:18:39.014772 4989 scope.go:117] "RemoveContainer" containerID="4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c" Dec 13 17:18:39 crc kubenswrapper[4989]: I1213 17:18:39.639998 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hllvq_1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf/kube-multus/1.log" Dec 13 17:18:39 crc kubenswrapper[4989]: I1213 17:18:39.640082 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hllvq" event={"ID":"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf","Type":"ContainerStarted","Data":"115cb6d4fb5c16108922c082c15d8ae751144a27bf4c118f70fbfceb4d6c3964"} Dec 13 17:18:40 crc kubenswrapper[4989]: I1213 17:18:40.013695 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:40 crc kubenswrapper[4989]: E1213 17:18:40.013838 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 13 17:18:40 crc kubenswrapper[4989]: I1213 17:18:40.013968 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:40 crc kubenswrapper[4989]: I1213 17:18:40.014041 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:40 crc kubenswrapper[4989]: I1213 17:18:40.014049 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:40 crc kubenswrapper[4989]: E1213 17:18:40.014140 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 13 17:18:40 crc kubenswrapper[4989]: E1213 17:18:40.014240 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lfpf8" podUID="7d912915-788e-412a-bae8-6eccd6b4c238" Dec 13 17:18:40 crc kubenswrapper[4989]: E1213 17:18:40.014357 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 13 17:18:42 crc kubenswrapper[4989]: I1213 17:18:42.014141 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:42 crc kubenswrapper[4989]: I1213 17:18:42.014242 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:18:42 crc kubenswrapper[4989]: I1213 17:18:42.014162 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:42 crc kubenswrapper[4989]: I1213 17:18:42.014492 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:42 crc kubenswrapper[4989]: I1213 17:18:42.018439 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 13 17:18:42 crc kubenswrapper[4989]: I1213 17:18:42.019760 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 13 17:18:42 crc kubenswrapper[4989]: I1213 17:18:42.020102 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 13 17:18:42 crc kubenswrapper[4989]: I1213 17:18:42.019774 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 13 17:18:42 crc kubenswrapper[4989]: I1213 17:18:42.020568 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 13 17:18:42 crc kubenswrapper[4989]: I1213 17:18:42.021534 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 13 17:18:48 crc kubenswrapper[4989]: I1213 17:18:48.408266 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.675032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.725493 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8hwgn"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.725916 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.728372 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5sdlr"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.728713 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.729944 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.730387 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.730834 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.730907 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.731178 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.731275 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.734596 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.735075 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.737308 4989 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-tls": failed to list *v1.Secret: secrets "machine-api-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.737398 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.737550 4989 reflector.go:561] object-"openshift-machine-api"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.737587 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.738056 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.738984 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739294 4989 reflector.go:561] object-"openshift-machine-api"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739329 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739337 4989 reflector.go:561] object-"openshift-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739377 4989 reflector.go:561] object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739416 4989 reflector.go:561] object-"openshift-machine-api"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739472 4989 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739464 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739386 4989 reflector.go:561] object-"openshift-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739486 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739471 4989 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7": failed to list *v1.Secret: secrets "machine-api-operator-dockercfg-mfbb7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739544 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739435 4989 reflector.go:561] object-"openshift-controller-manager"/"openshift-global-ca": failed to list *v1.ConfigMap: configmaps "openshift-global-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739560 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-mfbb7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-dockercfg-mfbb7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739565 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-global-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-global-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739423 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739384 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739588 4989 reflector.go:561] object-"openshift-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739613 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739628 4989 reflector.go:561] object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c": failed to list *v1.Secret: secrets "openshift-controller-manager-sa-dockercfg-msq4c" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739587 4989 reflector.go:561] object-"openshift-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739641 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-msq4c\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-controller-manager-sa-dockercfg-msq4c\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739651 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.739759 4989 reflector.go:561] object-"openshift-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.739892 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.740011 4989 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-images": failed to list *v1.ConfigMap: configmaps "machine-api-operator-images" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.740024 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-images\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"machine-api-operator-images\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.742879 4989 reflector.go:561] object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2": failed to list *v1.Secret: secrets "route-controller-manager-sa-dockercfg-h2zr2" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.742917 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-h2zr2\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"route-controller-manager-sa-dockercfg-h2zr2\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.742957 4989 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert": failed to list *v1.Secret: secrets "openshift-apiserver-operator-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.742970 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-operator-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.743073 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.743899 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.745409 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7sgxw"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.745878 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.748457 4989 reflector.go:561] object-"openshift-route-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.748500 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.748538 4989 reflector.go:561] object-"openshift-route-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.748549 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.748871 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.748987 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-drznw"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.749332 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-9kdlf"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.749484 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.749816 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.749907 4989 reflector.go:561] object-"openshift-apiserver-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.749933 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.750092 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.762283 4989 reflector.go:561] object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.762361 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.762461 4989 reflector.go:561] object-"openshift-cluster-samples-operator"/"samples-operator-tls": failed to list *v1.Secret: secrets "samples-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.762485 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"samples-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.762639 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.764408 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.764493 4989 reflector.go:561] object-"openshift-route-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.764609 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.764926 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.765025 4989 reflector.go:561] object-"openshift-authentication-operator"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.765093 4989 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv": failed to list *v1.Secret: secrets "openshift-apiserver-operator-dockercfg-xtcjv" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.765097 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.765125 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-dockercfg-xtcjv\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-operator-dockercfg-xtcjv\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.765387 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.765380 4989 reflector.go:561] object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z": failed to list *v1.Secret: secrets "openshift-config-operator-dockercfg-7pc5z" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-config-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.765448 4989 reflector.go:561] object-"openshift-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-config-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.765420 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-config-operator\"/\"openshift-config-operator-dockercfg-7pc5z\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-config-operator-dockercfg-7pc5z\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.765590 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.765475 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.765903 4989 reflector.go:561] object-"openshift-authentication-operator"/"authentication-operator-config": failed to list *v1.ConfigMap: configmaps "authentication-operator-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.765963 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.766016 4989 reflector.go:561] object-"openshift-route-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.781020 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.781985 4989 reflector.go:561] object-"openshift-authentication-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.782038 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.765969 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"authentication-operator-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"authentication-operator-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.782081 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.782111 4989 reflector.go:561] object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj": failed to list *v1.Secret: secrets "authentication-operator-dockercfg-mz9bj" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.782117 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.782191 4989 reflector.go:561] object-"openshift-authentication-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.782228 4989 reflector.go:561] object-"openshift-route-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.782224 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.782193 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.782281 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.782241 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.782313 4989 reflector.go:561] object-"openshift-authentication-operator"/"service-ca-bundle": failed to list *v1.ConfigMap: configmaps "service-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.782333 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"service-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"service-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.782348 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.782398 4989 reflector.go:561] object-"openshift-config-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-config-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.782405 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.782414 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-config-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.766129 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.782127 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"authentication-operator-dockercfg-mz9bj\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"authentication-operator-dockercfg-mz9bj\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.782288 4989 reflector.go:561] object-"openshift-config-operator"/"config-operator-serving-cert": failed to list *v1.Secret: secrets "config-operator-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-config-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.782546 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-config-operator\"/\"config-operator-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"config-operator-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.782576 4989 reflector.go:561] object-"openshift-cluster-samples-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.782586 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.783090 4989 reflector.go:561] object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w": failed to list *v1.Secret: secrets "cluster-samples-operator-dockercfg-xpp9w" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.784307 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-xpp9w\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cluster-samples-operator-dockercfg-xpp9w\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.784156 4989 reflector.go:561] object-"openshift-authentication-operator"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.784342 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.784199 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 13 17:18:49 crc kubenswrapper[4989]: W1213 17:18:49.784229 4989 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config": failed to list *v1.ConfigMap: configmaps "openshift-apiserver-operator-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Dec 13 17:18:49 crc kubenswrapper[4989]: E1213 17:18:49.784441 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-apiserver-operator-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.784540 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-lvjbs"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.784628 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.784730 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.784910 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.785188 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.785215 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.786338 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.786363 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.786655 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.786699 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-lvjbs" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.792397 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.792594 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dvtmv"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.793464 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.793507 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.793592 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.794176 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.794485 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.794528 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.797284 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.799207 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2c6kj"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.799286 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.800010 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.800544 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9blxx"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.801211 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sq6bg"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.801405 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.801554 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.801689 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.802234 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.802252 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.802449 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.802940 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.803255 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.803326 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.803455 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.803455 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.803920 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.804217 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.804346 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.807988 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.808212 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.808471 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.808588 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.808660 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.808748 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.808902 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.808939 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.808977 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.808904 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.809088 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.809149 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.809317 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.809460 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.809631 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-5t575"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.809676 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.809774 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.809898 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ncjdp"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.827010 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.809994 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.828828 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.829017 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.809976 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.810046 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.809961 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.810057 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.810092 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.810100 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.810211 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.810533 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.833296 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.833503 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.834236 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.834541 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.835145 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.835344 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.835521 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.836099 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k5w28"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.839689 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8hwgn"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.839996 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840290 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/96f26d01-772f-4e9c-b85b-d0585b936f8c-machine-approver-tls\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840331 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3be4123f-680d-4afa-a9a4-2a77100953c1-serving-cert\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840372 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840403 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af7c90c2-09e5-4f9d-8fa2-81dce6863688-serving-cert\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840547 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwwh7\" (UniqueName: \"kubernetes.io/projected/af7c90c2-09e5-4f9d-8fa2-81dce6863688-kube-api-access-qwwh7\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840569 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-config\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840589 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-config\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840608 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4f505d5a-020b-4a22-ae54-62bdfa3985f6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cxb7t\" (UID: \"4f505d5a-020b-4a22-ae54-62bdfa3985f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840623 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840638 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840654 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840685 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06291205-adcc-4352-9c00-ecd417c31f2a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pqk7b\" (UID: \"06291205-adcc-4352-9c00-ecd417c31f2a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840704 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840742 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htgjv\" (UniqueName: \"kubernetes.io/projected/96f26d01-772f-4e9c-b85b-d0585b936f8c-kube-api-access-htgjv\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840780 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgr2q\" (UniqueName: \"kubernetes.io/projected/3be4123f-680d-4afa-a9a4-2a77100953c1-kube-api-access-cgr2q\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840822 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96f26d01-772f-4e9c-b85b-d0585b936f8c-config\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840903 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgqzp\" (UniqueName: \"kubernetes.io/projected/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-kube-api-access-dgqzp\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840953 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-service-ca-bundle\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.840980 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmx5l\" (UniqueName: \"kubernetes.io/projected/03635dfa-6663-49c7-a980-9fda1f1f5a29-kube-api-access-qmx5l\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841003 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-images\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841044 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/96f26d01-772f-4e9c-b85b-d0585b936f8c-auth-proxy-config\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841082 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841104 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841125 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841147 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7kq2\" (UniqueName: \"kubernetes.io/projected/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-kube-api-access-f7kq2\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841191 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06291205-adcc-4352-9c00-ecd417c31f2a-config\") pod \"kube-controller-manager-operator-78b949d7b-pqk7b\" (UID: \"06291205-adcc-4352-9c00-ecd417c31f2a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841214 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f505d5a-020b-4a22-ae54-62bdfa3985f6-serving-cert\") pod \"openshift-config-operator-7777fb866f-cxb7t\" (UID: \"4f505d5a-020b-4a22-ae54-62bdfa3985f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841241 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxlbx\" (UniqueName: \"kubernetes.io/projected/4f505d5a-020b-4a22-ae54-62bdfa3985f6-kube-api-access-rxlbx\") pod \"openshift-config-operator-7777fb866f-cxb7t\" (UID: \"4f505d5a-020b-4a22-ae54-62bdfa3985f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841268 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skz4z\" (UniqueName: \"kubernetes.io/projected/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-kube-api-access-skz4z\") pod \"cluster-samples-operator-665b6dd947-94ggs\" (UID: \"b871b2b1-d492-4db5-91d8-3d77bcdd56e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841291 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-94ggs\" (UID: \"b871b2b1-d492-4db5-91d8-3d77bcdd56e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841317 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06291205-adcc-4352-9c00-ecd417c31f2a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pqk7b\" (UID: \"06291205-adcc-4352-9c00-ecd417c31f2a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841340 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.841364 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.842350 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.845717 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.845977 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.847984 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.849860 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.853075 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.859313 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9wfgs"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.859867 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.861739 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.862580 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.869131 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.869752 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.875229 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.876533 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.876564 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.877011 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.877389 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.877418 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.877574 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.877969 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5sdlr"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.877998 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.878093 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.878193 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.878258 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.878945 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.878972 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.879464 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.879578 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.879606 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kklpq"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.880099 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.880273 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.883542 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9wfgs"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.888758 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.890111 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.891885 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9blxx"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.898820 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dvtmv"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.898865 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-l68cn"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.899448 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-l68cn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.925366 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.926312 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.927165 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.927883 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.930912 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.933873 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2c6kj"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.936338 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.937086 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.938334 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-drznw"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.939266 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.941121 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.941700 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sq6bg"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942013 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af7c90c2-09e5-4f9d-8fa2-81dce6863688-serving-cert\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942410 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942439 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-etcd-ca\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942475 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-audit-dir\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942492 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942512 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwwh7\" (UniqueName: \"kubernetes.io/projected/af7c90c2-09e5-4f9d-8fa2-81dce6863688-kube-api-access-qwwh7\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942527 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-config\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942543 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942558 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-etcd-client\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942574 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c71f0f5d-4e9b-4a20-9007-22ec7b5e172e-proxy-tls\") pod \"machine-config-controller-84d6567774-qhk6m\" (UID: \"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942590 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk8cg\" (UniqueName: \"kubernetes.io/projected/70f58b07-ad87-4ea5-8373-4caae0199fdc-kube-api-access-dk8cg\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942607 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-config\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942623 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942638 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942654 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-config\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942671 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4f505d5a-020b-4a22-ae54-62bdfa3985f6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cxb7t\" (UID: \"4f505d5a-020b-4a22-ae54-62bdfa3985f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942686 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942702 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-etcd-serving-ca\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942716 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsbwz\" (UniqueName: \"kubernetes.io/projected/5cf82975-26bf-461d-af29-79560e79663a-kube-api-access-jsbwz\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942732 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942747 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942763 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-policies\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942800 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06291205-adcc-4352-9c00-ecd417c31f2a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pqk7b\" (UID: \"06291205-adcc-4352-9c00-ecd417c31f2a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942815 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942831 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d665e361-bdd1-4a87-a913-d67c2d15a102-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rmmll\" (UID: \"d665e361-bdd1-4a87-a913-d67c2d15a102\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942846 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18785359-0f0d-4177-a4d2-068af37b0730-serving-cert\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942859 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-trusted-ca\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942876 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htgjv\" (UniqueName: \"kubernetes.io/projected/96f26d01-772f-4e9c-b85b-d0585b936f8c-kube-api-access-htgjv\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942892 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgr2q\" (UniqueName: \"kubernetes.io/projected/3be4123f-680d-4afa-a9a4-2a77100953c1-kube-api-access-cgr2q\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.942906 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-dir\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943075 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943146 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96f26d01-772f-4e9c-b85b-d0585b936f8c-config\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943240 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-etcd-client\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943263 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-image-import-ca\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943307 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgqzp\" (UniqueName: \"kubernetes.io/projected/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-kube-api-access-dgqzp\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943324 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78a3473b-2599-43d9-98d5-83f2356618f5-config\") pod \"kube-apiserver-operator-766d6c64bb-656kn\" (UID: \"78a3473b-2599-43d9-98d5-83f2356618f5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943345 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-service-ca-bundle\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943464 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/18785359-0f0d-4177-a4d2-068af37b0730-node-pullsecrets\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943493 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k5w28\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943518 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmx5l\" (UniqueName: \"kubernetes.io/projected/03635dfa-6663-49c7-a980-9fda1f1f5a29-kube-api-access-qmx5l\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943538 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-images\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943560 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-audit-policies\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943642 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943667 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr8fr\" (UniqueName: \"kubernetes.io/projected/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-kube-api-access-rr8fr\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943672 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96f26d01-772f-4e9c-b85b-d0585b936f8c-config\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943686 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d665e361-bdd1-4a87-a913-d67c2d15a102-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rmmll\" (UID: \"d665e361-bdd1-4a87-a913-d67c2d15a102\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943737 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5cf82975-26bf-461d-af29-79560e79663a-proxy-tls\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943765 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4f505d5a-020b-4a22-ae54-62bdfa3985f6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cxb7t\" (UID: \"4f505d5a-020b-4a22-ae54-62bdfa3985f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943835 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/96f26d01-772f-4e9c-b85b-d0585b936f8c-auth-proxy-config\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943864 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943890 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/18785359-0f0d-4177-a4d2-068af37b0730-encryption-config\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943914 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-serving-cert\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.943936 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c71f0f5d-4e9b-4a20-9007-22ec7b5e172e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qhk6m\" (UID: \"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944081 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-service-ca\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944120 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78a3473b-2599-43d9-98d5-83f2356618f5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-656kn\" (UID: \"78a3473b-2599-43d9-98d5-83f2356618f5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944271 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944306 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zxtj\" (UniqueName: \"kubernetes.io/projected/d665e361-bdd1-4a87-a913-d67c2d15a102-kube-api-access-7zxtj\") pod \"openshift-controller-manager-operator-756b6f6bc6-rmmll\" (UID: \"d665e361-bdd1-4a87-a913-d67c2d15a102\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944329 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/18785359-0f0d-4177-a4d2-068af37b0730-audit-dir\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944348 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kvfv\" (UniqueName: \"kubernetes.io/projected/7b6a2091-2674-4874-a3aa-4b389b8be892-kube-api-access-8kvfv\") pod \"downloads-7954f5f757-lvjbs\" (UID: \"7b6a2091-2674-4874-a3aa-4b389b8be892\") " pod="openshift-console/downloads-7954f5f757-lvjbs" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944413 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgpc8\" (UniqueName: \"kubernetes.io/projected/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-kube-api-access-jgpc8\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944729 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7sgxw"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944757 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdkpp\" (UniqueName: \"kubernetes.io/projected/c71f0f5d-4e9b-4a20-9007-22ec7b5e172e-kube-api-access-qdkpp\") pod \"machine-config-controller-84d6567774-qhk6m\" (UID: \"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944810 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944830 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944851 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7kq2\" (UniqueName: \"kubernetes.io/projected/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-kube-api-access-f7kq2\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944872 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-etcd-service-ca\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944896 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-oauth-serving-cert\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944956 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06291205-adcc-4352-9c00-ecd417c31f2a-config\") pod \"kube-controller-manager-operator-78b949d7b-pqk7b\" (UID: \"06291205-adcc-4352-9c00-ecd417c31f2a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.944981 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-trusted-ca-bundle\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945009 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f505d5a-020b-4a22-ae54-62bdfa3985f6-serving-cert\") pod \"openshift-config-operator-7777fb866f-cxb7t\" (UID: \"4f505d5a-020b-4a22-ae54-62bdfa3985f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945031 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/18785359-0f0d-4177-a4d2-068af37b0730-etcd-client\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945054 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-serving-cert\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945075 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/70f58b07-ad87-4ea5-8373-4caae0199fdc-console-oauth-config\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945095 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78a3473b-2599-43d9-98d5-83f2356618f5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-656kn\" (UID: \"78a3473b-2599-43d9-98d5-83f2356618f5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945116 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csr8v\" (UniqueName: \"kubernetes.io/projected/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-kube-api-access-csr8v\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945143 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k5w28\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945162 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945181 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxlbx\" (UniqueName: \"kubernetes.io/projected/4f505d5a-020b-4a22-ae54-62bdfa3985f6-kube-api-access-rxlbx\") pod \"openshift-config-operator-7777fb866f-cxb7t\" (UID: \"4f505d5a-020b-4a22-ae54-62bdfa3985f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945199 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvhht\" (UniqueName: \"kubernetes.io/projected/18785359-0f0d-4177-a4d2-068af37b0730-kube-api-access-cvhht\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945215 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-console-config\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945232 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945250 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-config\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945265 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5cf82975-26bf-461d-af29-79560e79663a-images\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945282 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skz4z\" (UniqueName: \"kubernetes.io/projected/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-kube-api-access-skz4z\") pod \"cluster-samples-operator-665b6dd947-94ggs\" (UID: \"b871b2b1-d492-4db5-91d8-3d77bcdd56e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945298 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945315 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945332 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtd24\" (UniqueName: \"kubernetes.io/projected/bdf2d561-45dc-4672-929e-df9803b1d581-kube-api-access-wtd24\") pod \"marketplace-operator-79b997595-k5w28\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945347 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5njs\" (UniqueName: \"kubernetes.io/projected/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-kube-api-access-v5njs\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945367 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-94ggs\" (UID: \"b871b2b1-d492-4db5-91d8-3d77bcdd56e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945386 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06291205-adcc-4352-9c00-ecd417c31f2a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pqk7b\" (UID: \"06291205-adcc-4352-9c00-ecd417c31f2a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945405 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945423 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945493 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/96f26d01-772f-4e9c-b85b-d0585b936f8c-machine-approver-tls\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945511 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945527 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-encryption-config\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945544 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-audit\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945575 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945592 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3be4123f-680d-4afa-a9a4-2a77100953c1-serving-cert\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945609 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945625 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cf82975-26bf-461d-af29-79560e79663a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945641 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/70f58b07-ad87-4ea5-8373-4caae0199fdc-console-serving-cert\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945655 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-metrics-tls\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945674 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.945953 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06291205-adcc-4352-9c00-ecd417c31f2a-config\") pod \"kube-controller-manager-operator-78b949d7b-pqk7b\" (UID: \"06291205-adcc-4352-9c00-ecd417c31f2a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.946061 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.947612 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k5w28"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.947807 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/96f26d01-772f-4e9c-b85b-d0585b936f8c-auth-proxy-config\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.948473 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.949114 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06291205-adcc-4352-9c00-ecd417c31f2a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pqk7b\" (UID: \"06291205-adcc-4352-9c00-ecd417c31f2a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.952973 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9kdlf"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.953009 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.954921 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ncjdp"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.958892 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.963690 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.964366 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/96f26d01-772f-4e9c-b85b-d0585b936f8c-machine-approver-tls\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.965129 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.966143 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.967309 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.968347 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-lvjbs"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.969800 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-l68cn"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.970863 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.972002 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.972681 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.973136 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.974254 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xqhbq"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.975385 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-ghjq4"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.975486 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.976050 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.976620 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-ghjq4"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.977911 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kklpq"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.978903 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xqhbq"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.979927 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-j94xz"] Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.980610 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-j94xz" Dec 13 17:18:49 crc kubenswrapper[4989]: I1213 17:18:49.988074 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.010586 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.028400 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046469 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046500 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/18785359-0f0d-4177-a4d2-068af37b0730-encryption-config\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046520 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-serving-cert\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046538 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c71f0f5d-4e9b-4a20-9007-22ec7b5e172e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qhk6m\" (UID: \"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046557 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-service-ca\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046573 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78a3473b-2599-43d9-98d5-83f2356618f5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-656kn\" (UID: \"78a3473b-2599-43d9-98d5-83f2356618f5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046607 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zxtj\" (UniqueName: \"kubernetes.io/projected/d665e361-bdd1-4a87-a913-d67c2d15a102-kube-api-access-7zxtj\") pod \"openshift-controller-manager-operator-756b6f6bc6-rmmll\" (UID: \"d665e361-bdd1-4a87-a913-d67c2d15a102\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046623 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/18785359-0f0d-4177-a4d2-068af37b0730-audit-dir\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046638 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kvfv\" (UniqueName: \"kubernetes.io/projected/7b6a2091-2674-4874-a3aa-4b389b8be892-kube-api-access-8kvfv\") pod \"downloads-7954f5f757-lvjbs\" (UID: \"7b6a2091-2674-4874-a3aa-4b389b8be892\") " pod="openshift-console/downloads-7954f5f757-lvjbs" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046654 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgpc8\" (UniqueName: \"kubernetes.io/projected/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-kube-api-access-jgpc8\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046669 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdkpp\" (UniqueName: \"kubernetes.io/projected/c71f0f5d-4e9b-4a20-9007-22ec7b5e172e-kube-api-access-qdkpp\") pod \"machine-config-controller-84d6567774-qhk6m\" (UID: \"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046698 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-etcd-service-ca\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046712 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-oauth-serving-cert\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046730 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-trusted-ca-bundle\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046757 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/18785359-0f0d-4177-a4d2-068af37b0730-etcd-client\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046771 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csr8v\" (UniqueName: \"kubernetes.io/projected/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-kube-api-access-csr8v\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046806 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-serving-cert\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046822 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/70f58b07-ad87-4ea5-8373-4caae0199fdc-console-oauth-config\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046838 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78a3473b-2599-43d9-98d5-83f2356618f5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-656kn\" (UID: \"78a3473b-2599-43d9-98d5-83f2356618f5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046854 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k5w28\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046869 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046892 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvhht\" (UniqueName: \"kubernetes.io/projected/18785359-0f0d-4177-a4d2-068af37b0730-kube-api-access-cvhht\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046911 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-console-config\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046927 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046945 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-config\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046961 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5cf82975-26bf-461d-af29-79560e79663a-images\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046981 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046996 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047013 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtd24\" (UniqueName: \"kubernetes.io/projected/bdf2d561-45dc-4672-929e-df9803b1d581-kube-api-access-wtd24\") pod \"marketplace-operator-79b997595-k5w28\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047029 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5njs\" (UniqueName: \"kubernetes.io/projected/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-kube-api-access-v5njs\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047067 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047084 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-audit\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047099 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-encryption-config\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047141 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047155 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047171 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cf82975-26bf-461d-af29-79560e79663a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047185 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/70f58b07-ad87-4ea5-8373-4caae0199fdc-console-serving-cert\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047199 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-metrics-tls\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047229 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047245 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-etcd-ca\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047261 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-audit-dir\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047275 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047302 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047319 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-etcd-client\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047335 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c71f0f5d-4e9b-4a20-9007-22ec7b5e172e-proxy-tls\") pod \"machine-config-controller-84d6567774-qhk6m\" (UID: \"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047350 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk8cg\" (UniqueName: \"kubernetes.io/projected/70f58b07-ad87-4ea5-8373-4caae0199fdc-kube-api-access-dk8cg\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047371 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047388 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047404 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-config\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047424 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-etcd-serving-ca\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047440 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsbwz\" (UniqueName: \"kubernetes.io/projected/5cf82975-26bf-461d-af29-79560e79663a-kube-api-access-jsbwz\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047466 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-policies\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047470 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047493 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d665e361-bdd1-4a87-a913-d67c2d15a102-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rmmll\" (UID: \"d665e361-bdd1-4a87-a913-d67c2d15a102\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047510 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18785359-0f0d-4177-a4d2-068af37b0730-serving-cert\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047526 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-trusted-ca\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047550 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-dir\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047572 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047574 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c71f0f5d-4e9b-4a20-9007-22ec7b5e172e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qhk6m\" (UID: \"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047588 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-etcd-client\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047624 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-image-import-ca\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047656 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78a3473b-2599-43d9-98d5-83f2356618f5-config\") pod \"kube-apiserver-operator-766d6c64bb-656kn\" (UID: \"78a3473b-2599-43d9-98d5-83f2356618f5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047681 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/18785359-0f0d-4177-a4d2-068af37b0730-node-pullsecrets\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047701 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k5w28\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047732 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-audit-policies\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047740 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-service-ca\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047747 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047878 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr8fr\" (UniqueName: \"kubernetes.io/projected/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-kube-api-access-rr8fr\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047903 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d665e361-bdd1-4a87-a913-d67c2d15a102-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rmmll\" (UID: \"d665e361-bdd1-4a87-a913-d67c2d15a102\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.047921 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5cf82975-26bf-461d-af29-79560e79663a-proxy-tls\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.048340 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.048411 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/18785359-0f0d-4177-a4d2-068af37b0730-node-pullsecrets\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.048531 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.048679 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-image-import-ca\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.048780 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.048810 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-etcd-serving-ca\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.049153 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-console-config\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.046906 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/18785359-0f0d-4177-a4d2-068af37b0730-audit-dir\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.049268 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-policies\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.049327 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-dir\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.049521 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d665e361-bdd1-4a87-a913-d67c2d15a102-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rmmll\" (UID: \"d665e361-bdd1-4a87-a913-d67c2d15a102\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.050335 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-audit-policies\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.050550 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.050912 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/18785359-0f0d-4177-a4d2-068af37b0730-encryption-config\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.051002 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-audit-dir\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.051394 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.051507 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-oauth-serving-cert\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.051845 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-audit\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.051858 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.051896 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18785359-0f0d-4177-a4d2-068af37b0730-config\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.052080 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.052246 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cf82975-26bf-461d-af29-79560e79663a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.052492 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70f58b07-ad87-4ea5-8373-4caae0199fdc-trusted-ca-bundle\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.052727 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.052936 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-serving-cert\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.054074 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/70f58b07-ad87-4ea5-8373-4caae0199fdc-console-oauth-config\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.054162 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/70f58b07-ad87-4ea5-8373-4caae0199fdc-console-serving-cert\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.054849 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c71f0f5d-4e9b-4a20-9007-22ec7b5e172e-proxy-tls\") pod \"machine-config-controller-84d6567774-qhk6m\" (UID: \"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.056664 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.057055 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d665e361-bdd1-4a87-a913-d67c2d15a102-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rmmll\" (UID: \"d665e361-bdd1-4a87-a913-d67c2d15a102\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.057270 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.057470 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-encryption-config\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.057746 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.058110 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-etcd-client\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.066114 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.065922 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/18785359-0f0d-4177-a4d2-068af37b0730-etcd-client\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.066276 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.067622 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.067661 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.081249 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18785359-0f0d-4177-a4d2-068af37b0730-serving-cert\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.087355 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.107726 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.127485 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.147507 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.168403 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.187845 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.207264 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.227994 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.247433 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.268981 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.272462 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-metrics-tls\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.298256 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.300171 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-trusted-ca\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.307489 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.328311 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.347732 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.368016 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.380415 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-serving-cert\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.388204 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.399680 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-etcd-client\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.407856 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.428939 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.429640 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-config\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.447905 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.452497 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-etcd-ca\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.468606 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.479106 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-etcd-service-ca\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.488392 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.509263 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.527849 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.529513 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5cf82975-26bf-461d-af29-79560e79663a-images\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.548752 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.567545 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.587240 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.607833 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.612010 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5cf82975-26bf-461d-af29-79560e79663a-proxy-tls\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.627560 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.647855 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.668472 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.688227 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.708115 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.720517 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78a3473b-2599-43d9-98d5-83f2356618f5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-656kn\" (UID: \"78a3473b-2599-43d9-98d5-83f2356618f5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.727620 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.729734 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78a3473b-2599-43d9-98d5-83f2356618f5-config\") pod \"kube-apiserver-operator-766d6c64bb-656kn\" (UID: \"78a3473b-2599-43d9-98d5-83f2356618f5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.748363 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.768469 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.787291 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.807288 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.827138 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.846644 4989 request.go:700] Waited for 1.006103186s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.847672 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.867207 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.872020 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k5w28\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.887306 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.908132 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.934668 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.941316 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k5w28\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.943350 4989 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.943368 4989 secret.go:188] Couldn't get secret openshift-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.943396 4989 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.943439 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-machine-api-operator-tls podName:7fe4787f-cdd3-44fe-8405-dd6ebc9b455c nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.443402505 +0000 UTC m=+146.049849633 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-8hwgn" (UID: "7fe4787f-cdd3-44fe-8405-dd6ebc9b455c") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.943443 4989 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.943413 4989 configmap.go:193] Couldn't get configMap openshift-authentication-operator/authentication-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.943466 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af7c90c2-09e5-4f9d-8fa2-81dce6863688-serving-cert podName:af7c90c2-09e5-4f9d-8fa2-81dce6863688 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.443452747 +0000 UTC m=+146.049899985 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/af7c90c2-09e5-4f9d-8fa2-81dce6863688-serving-cert") pod "controller-manager-879f6c89f-5sdlr" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.943480 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert podName:03635dfa-6663-49c7-a980-9fda1f1f5a29 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.443473018 +0000 UTC m=+146.049920286 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert") pod "route-controller-manager-6576b87f9c-4wz97" (UID: "03635dfa-6663-49c7-a980-9fda1f1f5a29") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.943492 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-config podName:3be4123f-680d-4afa-a9a4-2a77100953c1 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.443486548 +0000 UTC m=+146.049933806 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-config") pod "authentication-operator-69f744f599-7sgxw" (UID: "3be4123f-680d-4afa-a9a4-2a77100953c1") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.943506 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca podName:03635dfa-6663-49c7-a980-9fda1f1f5a29 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.443498148 +0000 UTC m=+146.049945396 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca") pod "route-controller-manager-6576b87f9c-4wz97" (UID: "03635dfa-6663-49c7-a980-9fda1f1f5a29") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.944802 4989 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.944832 4989 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.944853 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config podName:03635dfa-6663-49c7-a980-9fda1f1f5a29 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.444837751 +0000 UTC m=+146.051284989 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config") pod "route-controller-manager-6576b87f9c-4wz97" (UID: "03635dfa-6663-49c7-a980-9fda1f1f5a29") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.944875 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-config podName:7fe4787f-cdd3-44fe-8405-dd6ebc9b455c nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.444859992 +0000 UTC m=+146.051307210 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-config") pod "machine-api-operator-5694c8668f-8hwgn" (UID: "7fe4787f-cdd3-44fe-8405-dd6ebc9b455c") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.944893 4989 configmap.go:193] Couldn't get configMap openshift-authentication-operator/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.944920 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-trusted-ca-bundle podName:3be4123f-680d-4afa-a9a4-2a77100953c1 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.444912313 +0000 UTC m=+146.051359451 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-trusted-ca-bundle") pod "authentication-operator-69f744f599-7sgxw" (UID: "3be4123f-680d-4afa-a9a4-2a77100953c1") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.946185 4989 secret.go:188] Couldn't get secret openshift-config-operator/config-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.946228 4989 secret.go:188] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.946209 4989 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.946256 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f505d5a-020b-4a22-ae54-62bdfa3985f6-serving-cert podName:4f505d5a-020b-4a22-ae54-62bdfa3985f6 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.446244815 +0000 UTC m=+146.052692043 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/4f505d5a-020b-4a22-ae54-62bdfa3985f6-serving-cert") pod "openshift-config-operator-7777fb866f-cxb7t" (UID: "4f505d5a-020b-4a22-ae54-62bdfa3985f6") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.946322 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-samples-operator-tls podName:b871b2b1-d492-4db5-91d8-3d77bcdd56e9 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.446301936 +0000 UTC m=+146.052749154 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-samples-operator-tls") pod "cluster-samples-operator-665b6dd947-94ggs" (UID: "b871b2b1-d492-4db5-91d8-3d77bcdd56e9") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.946235 4989 secret.go:188] Couldn't get secret openshift-apiserver-operator/openshift-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.946340 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca podName:af7c90c2-09e5-4f9d-8fa2-81dce6863688 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.446330847 +0000 UTC m=+146.052778115 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca") pod "controller-manager-879f6c89f-5sdlr" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.946380 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-serving-cert podName:fc182b3d-822b-44aa-a1ba-0af7c2ba5557 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.446363658 +0000 UTC m=+146.052810796 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-serving-cert") pod "openshift-apiserver-operator-796bbdcf4f-kf2q8" (UID: "fc182b3d-822b-44aa-a1ba-0af7c2ba5557") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.946410 4989 configmap.go:193] Couldn't get configMap openshift-apiserver-operator/openshift-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.946446 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-config podName:fc182b3d-822b-44aa-a1ba-0af7c2ba5557 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.446435831 +0000 UTC m=+146.052882969 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-config") pod "openshift-apiserver-operator-796bbdcf4f-kf2q8" (UID: "fc182b3d-822b-44aa-a1ba-0af7c2ba5557") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.947026 4989 configmap.go:193] Couldn't get configMap openshift-authentication-operator/service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.947058 4989 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.947084 4989 secret.go:188] Couldn't get secret openshift-authentication-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.947101 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-service-ca-bundle podName:3be4123f-680d-4afa-a9a4-2a77100953c1 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.447085421 +0000 UTC m=+146.053532649 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-service-ca-bundle") pod "authentication-operator-69f744f599-7sgxw" (UID: "3be4123f-680d-4afa-a9a4-2a77100953c1") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.947146 4989 configmap.go:193] Couldn't get configMap openshift-controller-manager/openshift-global-ca: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.947151 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config podName:af7c90c2-09e5-4f9d-8fa2-81dce6863688 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.447141763 +0000 UTC m=+146.053588901 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config") pod "controller-manager-879f6c89f-5sdlr" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.947166 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3be4123f-680d-4afa-a9a4-2a77100953c1-serving-cert podName:3be4123f-680d-4afa-a9a4-2a77100953c1 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.447159734 +0000 UTC m=+146.053606872 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/3be4123f-680d-4afa-a9a4-2a77100953c1-serving-cert") pod "authentication-operator-69f744f599-7sgxw" (UID: "3be4123f-680d-4afa-a9a4-2a77100953c1") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.947180 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles podName:af7c90c2-09e5-4f9d-8fa2-81dce6863688 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.447173824 +0000 UTC m=+146.053620962 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-ca-bundles" (UniqueName: "kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles") pod "controller-manager-879f6c89f-5sdlr" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.948032 4989 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: E1213 17:18:50.948080 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-images podName:7fe4787f-cdd3-44fe-8405-dd6ebc9b455c nodeName:}" failed. No retries permitted until 2025-12-13 17:18:51.448067563 +0000 UTC m=+146.054514791 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-images") pod "machine-api-operator-5694c8668f-8hwgn" (UID: "7fe4787f-cdd3-44fe-8405-dd6ebc9b455c") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:50 crc kubenswrapper[4989]: I1213 17:18:50.988198 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.007740 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.027736 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.048319 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.068405 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.087346 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.108846 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.127899 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.147687 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.168563 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.190174 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.208077 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.228342 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.248300 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.268044 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.287420 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.307427 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.328550 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.347894 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.367600 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.388089 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.407689 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.427388 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.448533 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.468434 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-service-ca-bundle\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.468487 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.468512 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-images\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.468617 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.468713 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.468759 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.468839 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f505d5a-020b-4a22-ae54-62bdfa3985f6-serving-cert\") pod \"openshift-config-operator-7777fb866f-cxb7t\" (UID: \"4f505d5a-020b-4a22-ae54-62bdfa3985f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.468960 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-94ggs\" (UID: \"b871b2b1-d492-4db5-91d8-3d77bcdd56e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.469007 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.469042 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.469115 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.469200 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3be4123f-680d-4afa-a9a4-2a77100953c1-serving-cert\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.469297 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af7c90c2-09e5-4f9d-8fa2-81dce6863688-serving-cert\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.469353 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-config\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.469386 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-config\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.469432 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.469464 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.469498 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.469545 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.488620 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.509016 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.527559 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.548639 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.568685 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.588335 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.725222 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06291205-adcc-4352-9c00-ecd417c31f2a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pqk7b\" (UID: \"06291205-adcc-4352-9c00-ecd417c31f2a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.788119 4989 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.808897 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.827687 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.847856 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.866892 4989 request.go:700] Waited for 1.890692856s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-default-metrics-tls&limit=500&resourceVersion=0 Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.868713 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.888522 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.907509 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.907765 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.928439 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.948068 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.975155 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:51 crc kubenswrapper[4989]: E1213 17:18:51.975350 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:20:53.975316552 +0000 UTC m=+268.581763700 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:51 crc kubenswrapper[4989]: I1213 17:18:51.990233 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zxtj\" (UniqueName: \"kubernetes.io/projected/d665e361-bdd1-4a87-a913-d67c2d15a102-kube-api-access-7zxtj\") pod \"openshift-controller-manager-operator-756b6f6bc6-rmmll\" (UID: \"d665e361-bdd1-4a87-a913-d67c2d15a102\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.010551 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kvfv\" (UniqueName: \"kubernetes.io/projected/7b6a2091-2674-4874-a3aa-4b389b8be892-kube-api-access-8kvfv\") pod \"downloads-7954f5f757-lvjbs\" (UID: \"7b6a2091-2674-4874-a3aa-4b389b8be892\") " pod="openshift-console/downloads-7954f5f757-lvjbs" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.026945 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgpc8\" (UniqueName: \"kubernetes.io/projected/3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3-kube-api-access-jgpc8\") pod \"apiserver-7bbb656c7d-sjxxg\" (UID: \"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.036547 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-lvjbs" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.049582 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csr8v\" (UniqueName: \"kubernetes.io/projected/1b4f9f0c-2b37-4dbe-9dff-ae118b994802-kube-api-access-csr8v\") pod \"etcd-operator-b45778765-ncjdp\" (UID: \"1b4f9f0c-2b37-4dbe-9dff-ae118b994802\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.062960 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.072307 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.077166 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.077446 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.077491 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.077579 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.077644 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.079019 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.080348 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.081334 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.082347 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.085658 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdkpp\" (UniqueName: \"kubernetes.io/projected/c71f0f5d-4e9b-4a20-9007-22ec7b5e172e-kube-api-access-qdkpp\") pod \"machine-config-controller-84d6567774-qhk6m\" (UID: \"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.096905 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.102778 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvhht\" (UniqueName: \"kubernetes.io/projected/18785359-0f0d-4177-a4d2-068af37b0730-kube-api-access-cvhht\") pod \"apiserver-76f77b778f-2c6kj\" (UID: \"18785359-0f0d-4177-a4d2-068af37b0730\") " pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.124139 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr8fr\" (UniqueName: \"kubernetes.io/projected/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-kube-api-access-rr8fr\") pod \"oauth-openshift-558db77b4-dvtmv\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.142566 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.146179 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsbwz\" (UniqueName: \"kubernetes.io/projected/5cf82975-26bf-461d-af29-79560e79663a-kube-api-access-jsbwz\") pod \"machine-config-operator-74547568cd-v97j6\" (UID: \"5cf82975-26bf-461d-af29-79560e79663a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.146404 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.165114 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtd24\" (UniqueName: \"kubernetes.io/projected/bdf2d561-45dc-4672-929e-df9803b1d581-kube-api-access-wtd24\") pod \"marketplace-operator-79b997595-k5w28\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.184900 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5njs\" (UniqueName: \"kubernetes.io/projected/04b35497-4fb1-4eae-a6d8-128c5f9aaf5c-kube-api-access-v5njs\") pod \"ingress-operator-5b745b69d9-7x4jv\" (UID: \"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.200883 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.204294 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78a3473b-2599-43d9-98d5-83f2356618f5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-656kn\" (UID: \"78a3473b-2599-43d9-98d5-83f2356618f5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.223744 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk8cg\" (UniqueName: \"kubernetes.io/projected/70f58b07-ad87-4ea5-8373-4caae0199fdc-kube-api-access-dk8cg\") pod \"console-f9d7485db-9kdlf\" (UID: \"70f58b07-ad87-4ea5-8373-4caae0199fdc\") " pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.228062 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.244489 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.268274 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.271396 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.277435 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af7c90c2-09e5-4f9d-8fa2-81dce6863688-serving-cert\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.283116 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.289977 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.292678 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.294256 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.310907 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.318309 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.324764 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-lvjbs"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.337554 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.345872 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.346172 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.347650 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.355754 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.357858 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7kq2\" (UniqueName: \"kubernetes.io/projected/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-kube-api-access-f7kq2\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:52 crc kubenswrapper[4989]: W1213 17:18:52.360239 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06291205_adcc_4352_9c00_ecd417c31f2a.slice/crio-9052abfb8531a1ea5af9be5fe1327520545b7ca6d67c547a10f4dcc91248af11 WatchSource:0}: Error finding container 9052abfb8531a1ea5af9be5fe1327520545b7ca6d67c547a10f4dcc91248af11: Status 404 returned error can't find the container with id 9052abfb8531a1ea5af9be5fe1327520545b7ca6d67c547a10f4dcc91248af11 Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.371068 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.379555 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-service-ca-bundle\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.383835 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.388482 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.390939 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-config\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:52 crc kubenswrapper[4989]: W1213 17:18:52.399524 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd665e361_bdd1_4a87_a913_d67c2d15a102.slice/crio-311f524c7e1ba7bfb243828f61e9eb3c458af39a34e8c1e8cbaf444190a00413 WatchSource:0}: Error finding container 311f524c7e1ba7bfb243828f61e9eb3c458af39a34e8c1e8cbaf444190a00413: Status 404 returned error can't find the container with id 311f524c7e1ba7bfb243828f61e9eb3c458af39a34e8c1e8cbaf444190a00413 Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.407987 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.424141 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.428584 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.437805 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ncjdp"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.439521 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skz4z\" (UniqueName: \"kubernetes.io/projected/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-kube-api-access-skz4z\") pod \"cluster-samples-operator-665b6dd947-94ggs\" (UID: \"b871b2b1-d492-4db5-91d8-3d77bcdd56e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.448175 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469199 4989 secret.go:188] Couldn't get secret openshift-config-operator/config-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469272 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f505d5a-020b-4a22-ae54-62bdfa3985f6-serving-cert podName:4f505d5a-020b-4a22-ae54-62bdfa3985f6 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.469250828 +0000 UTC m=+148.075697966 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/4f505d5a-020b-4a22-ae54-62bdfa3985f6-serving-cert") pod "openshift-config-operator-7777fb866f-cxb7t" (UID: "4f505d5a-020b-4a22-ae54-62bdfa3985f6") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469318 4989 configmap.go:193] Couldn't get configMap openshift-controller-manager/openshift-global-ca: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469349 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles podName:af7c90c2-09e5-4f9d-8fa2-81dce6863688 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.469338561 +0000 UTC m=+148.075785699 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "proxy-ca-bundles" (UniqueName: "kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles") pod "controller-manager-879f6c89f-5sdlr" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469372 4989 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469401 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-images podName:7fe4787f-cdd3-44fe-8405-dd6ebc9b455c nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.469392773 +0000 UTC m=+148.075839911 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-images") pod "machine-api-operator-5694c8668f-8hwgn" (UID: "7fe4787f-cdd3-44fe-8405-dd6ebc9b455c") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469419 4989 secret.go:188] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469446 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-samples-operator-tls podName:b871b2b1-d492-4db5-91d8-3d77bcdd56e9 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.469439114 +0000 UTC m=+148.075886252 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-samples-operator-tls") pod "cluster-samples-operator-665b6dd947-94ggs" (UID: "b871b2b1-d492-4db5-91d8-3d77bcdd56e9") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469471 4989 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469495 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca podName:af7c90c2-09e5-4f9d-8fa2-81dce6863688 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.469487596 +0000 UTC m=+148.075934734 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca") pod "controller-manager-879f6c89f-5sdlr" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.469515 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469526 4989 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469549 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config podName:03635dfa-6663-49c7-a980-9fda1f1f5a29 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.469542688 +0000 UTC m=+148.075989826 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config") pod "route-controller-manager-6576b87f9c-4wz97" (UID: "03635dfa-6663-49c7-a980-9fda1f1f5a29") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469572 4989 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469593 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config podName:af7c90c2-09e5-4f9d-8fa2-81dce6863688 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.469586669 +0000 UTC m=+148.076033807 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config") pod "controller-manager-879f6c89f-5sdlr" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469815 4989 configmap.go:193] Couldn't get configMap openshift-apiserver-operator/openshift-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469831 4989 configmap.go:193] Couldn't get configMap openshift-authentication-operator/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469856 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-config podName:fc182b3d-822b-44aa-a1ba-0af7c2ba5557 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.469843757 +0000 UTC m=+148.076290895 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-config") pod "openshift-apiserver-operator-796bbdcf4f-kf2q8" (UID: "fc182b3d-822b-44aa-a1ba-0af7c2ba5557") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469864 4989 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469872 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-trusted-ca-bundle podName:3be4123f-680d-4afa-a9a4-2a77100953c1 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.469864288 +0000 UTC m=+148.076311426 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-trusted-ca-bundle") pod "authentication-operator-69f744f599-7sgxw" (UID: "3be4123f-680d-4afa-a9a4-2a77100953c1") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469893 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-machine-api-operator-tls podName:7fe4787f-cdd3-44fe-8405-dd6ebc9b455c nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.469884479 +0000 UTC m=+148.076331617 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-8hwgn" (UID: "7fe4787f-cdd3-44fe-8405-dd6ebc9b455c") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469926 4989 secret.go:188] Couldn't get secret openshift-authentication-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.469992 4989 configmap.go:193] Couldn't get configMap openshift-authentication-operator/authentication-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.470010 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3be4123f-680d-4afa-a9a4-2a77100953c1-serving-cert podName:3be4123f-680d-4afa-a9a4-2a77100953c1 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.469967041 +0000 UTC m=+148.076414179 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/3be4123f-680d-4afa-a9a4-2a77100953c1-serving-cert") pod "authentication-operator-69f744f599-7sgxw" (UID: "3be4123f-680d-4afa-a9a4-2a77100953c1") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.470032 4989 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.470032 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-config podName:3be4123f-680d-4afa-a9a4-2a77100953c1 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.470023023 +0000 UTC m=+148.076470161 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-config") pod "authentication-operator-69f744f599-7sgxw" (UID: "3be4123f-680d-4afa-a9a4-2a77100953c1") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.470068 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca podName:03635dfa-6663-49c7-a980-9fda1f1f5a29 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.470058424 +0000 UTC m=+148.076505562 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca") pod "route-controller-manager-6576b87f9c-4wz97" (UID: "03635dfa-6663-49c7-a980-9fda1f1f5a29") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.470094 4989 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.470157 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert podName:03635dfa-6663-49c7-a980-9fda1f1f5a29 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.470149987 +0000 UTC m=+148.076597125 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert") pod "route-controller-manager-6576b87f9c-4wz97" (UID: "03635dfa-6663-49c7-a980-9fda1f1f5a29") : failed to sync secret cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.485573 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.491633 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.493734 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.508564 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmx5l\" (UniqueName: \"kubernetes.io/projected/03635dfa-6663-49c7-a980-9fda1f1f5a29-kube-api-access-qmx5l\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.509476 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.528654 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.549281 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.559989 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k5w28"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.581373 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.587258 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.596530 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.602064 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.607767 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.671007 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.679080 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxlbx\" (UniqueName: \"kubernetes.io/projected/4f505d5a-020b-4a22-ae54-62bdfa3985f6-kube-api-access-rxlbx\") pod \"openshift-config-operator-7777fb866f-cxb7t\" (UID: \"4f505d5a-020b-4a22-ae54-62bdfa3985f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.682101 4989 projected.go:288] Couldn't get configMap openshift-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.689193 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.715989 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.729284 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.733162 4989 projected.go:194] Error preparing data for projected volume kube-api-access-qwwh7 for pod openshift-controller-manager/controller-manager-879f6c89f-5sdlr: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.733247 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/af7c90c2-09e5-4f9d-8fa2-81dce6863688-kube-api-access-qwwh7 podName:af7c90c2-09e5-4f9d-8fa2-81dce6863688 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.233221836 +0000 UTC m=+147.839668984 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-qwwh7" (UniqueName: "kubernetes.io/projected/af7c90c2-09e5-4f9d-8fa2-81dce6863688-kube-api-access-qwwh7") pod "controller-manager-879f6c89f-5sdlr" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.744303 4989 projected.go:288] Couldn't get configMap openshift-cluster-machine-approver/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.744352 4989 projected.go:194] Error preparing data for projected volume kube-api-access-htgjv for pod openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.744427 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/96f26d01-772f-4e9c-b85b-d0585b936f8c-kube-api-access-htgjv podName:96f26d01-772f-4e9c-b85b-d0585b936f8c nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.244388137 +0000 UTC m=+147.850835275 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-htgjv" (UniqueName: "kubernetes.io/projected/96f26d01-772f-4e9c-b85b-d0585b936f8c-kube-api-access-htgjv") pod "machine-approver-56656f9798-c7m2r" (UID: "96f26d01-772f-4e9c-b85b-d0585b936f8c") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.748186 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.765961 4989 projected.go:288] Couldn't get configMap openshift-apiserver-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.766015 4989 projected.go:194] Error preparing data for projected volume kube-api-access-dgqzp for pod openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8: failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: E1213 17:18:52.766081 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-kube-api-access-dgqzp podName:fc182b3d-822b-44aa-a1ba-0af7c2ba5557 nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.266062597 +0000 UTC m=+147.872509725 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-dgqzp" (UniqueName: "kubernetes.io/projected/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-kube-api-access-dgqzp") pod "openshift-apiserver-operator-796bbdcf4f-kf2q8" (UID: "fc182b3d-822b-44aa-a1ba-0af7c2ba5557") : failed to sync configmap cache: timed out waiting for the condition Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.767876 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.771887 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.788636 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.808662 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.828189 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.848567 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.868050 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.887029 4989 request.go:700] Waited for 1.695177869s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/configmaps?fieldSelector=metadata.name%3Dconfig&limit=500&resourceVersion=0 Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.888989 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.908087 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.921976 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2c6kj"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.927958 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.931172 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9kdlf"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.932511 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dvtmv"] Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.948174 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.967679 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 13 17:18:52 crc kubenswrapper[4989]: I1213 17:18:52.988161 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.008475 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.028467 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.038352 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ltx5\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-kube-api-access-6ltx5\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.038406 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-bound-sa-token\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.038433 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.038452 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-certificates\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.038504 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-tls\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.038540 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-trusted-ca\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.038566 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.038602 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: E1213 17:18:53.039688 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.539673246 +0000 UTC m=+148.146120384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:53 crc kubenswrapper[4989]: W1213 17:18:53.046674 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cf82975_26bf_461d_af29_79560e79663a.slice/crio-e42edee76be47d3e530262bce42cbe2d7c584e1da71dc060b9c444356d2410c6 WatchSource:0}: Error finding container e42edee76be47d3e530262bce42cbe2d7c584e1da71dc060b9c444356d2410c6: Status 404 returned error can't find the container with id e42edee76be47d3e530262bce42cbe2d7c584e1da71dc060b9c444356d2410c6 Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.048929 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.057941 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgr2q\" (UniqueName: \"kubernetes.io/projected/3be4123f-680d-4afa-a9a4-2a77100953c1-kube-api-access-cgr2q\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:53 crc kubenswrapper[4989]: W1213 17:18:53.061430 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-49066fafeb22adadee45414264bbf0d67999cdbf94ff31ecb7dceb90efd2239f WatchSource:0}: Error finding container 49066fafeb22adadee45414264bbf0d67999cdbf94ff31ecb7dceb90efd2239f: Status 404 returned error can't find the container with id 49066fafeb22adadee45414264bbf0d67999cdbf94ff31ecb7dceb90efd2239f Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.064272 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lvjbs" event={"ID":"7b6a2091-2674-4874-a3aa-4b389b8be892","Type":"ContainerStarted","Data":"619a89028cefed58130fadfe313d851417f292d743dc19dfdc645fc8a8675177"} Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.065302 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" event={"ID":"06291205-adcc-4352-9c00-ecd417c31f2a","Type":"ContainerStarted","Data":"9052abfb8531a1ea5af9be5fe1327520545b7ca6d67c547a10f4dcc91248af11"} Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.066106 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" event={"ID":"d665e361-bdd1-4a87-a913-d67c2d15a102","Type":"ContainerStarted","Data":"311f524c7e1ba7bfb243828f61e9eb3c458af39a34e8c1e8cbaf444190a00413"} Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.069121 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" event={"ID":"5cf82975-26bf-461d-af29-79560e79663a","Type":"ContainerStarted","Data":"e42edee76be47d3e530262bce42cbe2d7c584e1da71dc060b9c444356d2410c6"} Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.069835 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" event={"ID":"1b4f9f0c-2b37-4dbe-9dff-ae118b994802","Type":"ContainerStarted","Data":"6b6929adda0d011dd8e1adc828e047c87c71143888204ec44e6c6389d28ee3d3"} Dec 13 17:18:53 crc kubenswrapper[4989]: W1213 17:18:53.074759 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04b35497_4fb1_4eae_a6d8_128c5f9aaf5c.slice/crio-8394a18d7cce7d0dda062f88ff4cdf3a773e22f81e942d78020f980f28364da0 WatchSource:0}: Error finding container 8394a18d7cce7d0dda062f88ff4cdf3a773e22f81e942d78020f980f28364da0: Status 404 returned error can't find the container with id 8394a18d7cce7d0dda062f88ff4cdf3a773e22f81e942d78020f980f28364da0 Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.139218 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.139737 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-bound-sa-token\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.139873 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bj4fz\" (UID: \"fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.139913 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d96m\" (UniqueName: \"kubernetes.io/projected/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-kube-api-access-5d96m\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.139958 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-trusted-ca\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.139988 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3471f4e0-bf54-427b-9a0b-3b3242ff81e5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vjmm\" (UID: \"3471f4e0-bf54-427b-9a0b-3b3242ff81e5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.140023 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/10180d3c-e16a-46d0-a585-63f24d8587cd-metrics-tls\") pod \"dns-operator-744455d44c-sq6bg\" (UID: \"10180d3c-e16a-46d0-a585-63f24d8587cd\") " pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.140046 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72170a8d-7210-4a60-9ffc-9871a545fe36-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.140069 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3471f4e0-bf54-427b-9a0b-3b3242ff81e5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vjmm\" (UID: \"3471f4e0-bf54-427b-9a0b-3b3242ff81e5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.140093 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh64p\" (UniqueName: \"kubernetes.io/projected/72170a8d-7210-4a60-9ffc-9871a545fe36-kube-api-access-nh64p\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.140113 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-trusted-ca\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.140136 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a33960ce-7433-4ae4-a59f-df7a21d886d6-default-certificate\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.140156 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a33960ce-7433-4ae4-a59f-df7a21d886d6-service-ca-bundle\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.140615 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.140965 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: E1213 17:18:53.140998 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.640978647 +0000 UTC m=+148.247425795 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141026 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-certificates\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141057 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67vrt\" (UniqueName: \"kubernetes.io/projected/10180d3c-e16a-46d0-a585-63f24d8587cd-kube-api-access-67vrt\") pod \"dns-operator-744455d44c-sq6bg\" (UID: \"10180d3c-e16a-46d0-a585-63f24d8587cd\") " pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141085 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/72170a8d-7210-4a60-9ffc-9871a545fe36-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141111 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-tls\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141135 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a33960ce-7433-4ae4-a59f-df7a21d886d6-metrics-certs\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141160 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bppw\" (UniqueName: \"kubernetes.io/projected/a33960ce-7433-4ae4-a59f-df7a21d886d6-kube-api-access-2bppw\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141199 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141222 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-config\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141247 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bj4fz\" (UID: \"fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141273 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndbfp\" (UniqueName: \"kubernetes.io/projected/3471f4e0-bf54-427b-9a0b-3b3242ff81e5-kube-api-access-ndbfp\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vjmm\" (UID: \"3471f4e0-bf54-427b-9a0b-3b3242ff81e5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141323 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141353 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-serving-cert\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141380 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bj4fz\" (UID: \"fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141411 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/72170a8d-7210-4a60-9ffc-9871a545fe36-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141434 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a33960ce-7433-4ae4-a59f-df7a21d886d6-stats-auth\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.141518 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ltx5\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-kube-api-access-6ltx5\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.142001 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-certificates\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: E1213 17:18:53.142594 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.642580148 +0000 UTC m=+148.249027286 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.142601 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-trusted-ca\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.151040 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.152144 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-tls\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.186708 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-bound-sa-token\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.198662 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ltx5\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-kube-api-access-6ltx5\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.242604 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:53 crc kubenswrapper[4989]: E1213 17:18:53.242864 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.742801714 +0000 UTC m=+148.349248852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243241 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4aed05ac-1200-48f7-b840-dc71374cbb51-config\") pod \"service-ca-operator-777779d784-9bj6c\" (UID: \"4aed05ac-1200-48f7-b840-dc71374cbb51\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243314 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4hmv\" (UniqueName: \"kubernetes.io/projected/cbfd74ca-ade6-4943-95f6-6f930b2d571b-kube-api-access-x4hmv\") pod \"dns-default-ghjq4\" (UID: \"cbfd74ca-ade6-4943-95f6-6f930b2d571b\") " pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243339 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7ee8c55e-83e9-4786-a978-60457031890f-signing-cabundle\") pod \"service-ca-9c57cc56f-kklpq\" (UID: \"7ee8c55e-83e9-4786-a978-60457031890f\") " pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243364 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3471f4e0-bf54-427b-9a0b-3b3242ff81e5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vjmm\" (UID: \"3471f4e0-bf54-427b-9a0b-3b3242ff81e5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243387 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmr4p\" (UniqueName: \"kubernetes.io/projected/70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0-kube-api-access-rmr4p\") pod \"catalog-operator-68c6474976-7ssm5\" (UID: \"70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243440 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-registration-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243462 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72170a8d-7210-4a60-9ffc-9871a545fe36-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243480 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/10180d3c-e16a-46d0-a585-63f24d8587cd-metrics-tls\") pod \"dns-operator-744455d44c-sq6bg\" (UID: \"10180d3c-e16a-46d0-a585-63f24d8587cd\") " pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243514 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3471f4e0-bf54-427b-9a0b-3b3242ff81e5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vjmm\" (UID: \"3471f4e0-bf54-427b-9a0b-3b3242ff81e5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243553 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh64p\" (UniqueName: \"kubernetes.io/projected/72170a8d-7210-4a60-9ffc-9871a545fe36-kube-api-access-nh64p\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243576 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-trusted-ca\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243612 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6h6c\" (UniqueName: \"kubernetes.io/projected/7fb1e955-3966-463a-9196-0f9b8ce0f834-kube-api-access-z6h6c\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243638 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a33960ce-7433-4ae4-a59f-df7a21d886d6-default-certificate\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243656 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a33960ce-7433-4ae4-a59f-df7a21d886d6-service-ca-bundle\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243671 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-mountpoint-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243734 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r72x6\" (UniqueName: \"kubernetes.io/projected/0607f893-8d28-4498-8ac2-84776eaa26f2-kube-api-access-r72x6\") pod \"migrator-59844c95c7-qbf8q\" (UID: \"0607f893-8d28-4498-8ac2-84776eaa26f2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243751 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/60e9b65b-18c2-45a6-8295-d2dac4a939cd-certs\") pod \"machine-config-server-j94xz\" (UID: \"60e9b65b-18c2-45a6-8295-d2dac4a939cd\") " pod="openshift-machine-config-operator/machine-config-server-j94xz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243867 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7ee8c55e-83e9-4786-a978-60457031890f-signing-key\") pod \"service-ca-9c57cc56f-kklpq\" (UID: \"7ee8c55e-83e9-4786-a978-60457031890f\") " pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243913 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4b2g\" (UniqueName: \"kubernetes.io/projected/056cfe82-16b2-4c97-bbda-617a10007162-kube-api-access-h4b2g\") pod \"package-server-manager-789f6589d5-4pf8r\" (UID: \"056cfe82-16b2-4c97-bbda-617a10007162\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.243987 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5f63c6f4-9a0c-494e-966d-ccd469dc56c3-srv-cert\") pod \"olm-operator-6b444d44fb-l7htm\" (UID: \"5f63c6f4-9a0c-494e-966d-ccd469dc56c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244008 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7abd827d-5201-4f95-95eb-5a817ff7da69-tmpfs\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244030 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz9hv\" (UniqueName: \"kubernetes.io/projected/7abd827d-5201-4f95-95eb-5a817ff7da69-kube-api-access-gz9hv\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244074 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzg48\" (UniqueName: \"kubernetes.io/projected/f960de2c-fee1-4571-a4b4-5f47d796bd88-kube-api-access-fzg48\") pod \"multus-admission-controller-857f4d67dd-9wfgs\" (UID: \"f960de2c-fee1-4571-a4b4-5f47d796bd88\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244097 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5f63c6f4-9a0c-494e-966d-ccd469dc56c3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-l7htm\" (UID: \"5f63c6f4-9a0c-494e-966d-ccd469dc56c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244135 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67vrt\" (UniqueName: \"kubernetes.io/projected/10180d3c-e16a-46d0-a585-63f24d8587cd-kube-api-access-67vrt\") pod \"dns-operator-744455d44c-sq6bg\" (UID: \"10180d3c-e16a-46d0-a585-63f24d8587cd\") " pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244169 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/72170a8d-7210-4a60-9ffc-9871a545fe36-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244218 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a33960ce-7433-4ae4-a59f-df7a21d886d6-metrics-certs\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244253 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bppw\" (UniqueName: \"kubernetes.io/projected/a33960ce-7433-4ae4-a59f-df7a21d886d6-kube-api-access-2bppw\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244317 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0-profile-collector-cert\") pod \"catalog-operator-68c6474976-7ssm5\" (UID: \"70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244355 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a134b88a-8fa0-486b-b66d-b1df520f4a6c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fkvxp\" (UID: \"a134b88a-8fa0-486b-b66d-b1df520f4a6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244393 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-config\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244408 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7abd827d-5201-4f95-95eb-5a817ff7da69-webhook-cert\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.244453 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bj4fz\" (UID: \"fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.246215 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-trusted-ca\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.247293 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3471f4e0-bf54-427b-9a0b-3b3242ff81e5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vjmm\" (UID: \"3471f4e0-bf54-427b-9a0b-3b3242ff81e5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.248642 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a33960ce-7433-4ae4-a59f-df7a21d886d6-service-ca-bundle\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.253637 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/10180d3c-e16a-46d0-a585-63f24d8587cd-metrics-tls\") pod \"dns-operator-744455d44c-sq6bg\" (UID: \"10180d3c-e16a-46d0-a585-63f24d8587cd\") " pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.254461 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a33960ce-7433-4ae4-a59f-df7a21d886d6-default-certificate\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.257157 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/72170a8d-7210-4a60-9ffc-9871a545fe36-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.257872 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72170a8d-7210-4a60-9ffc-9871a545fe36-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.258024 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-config\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.261459 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndbfp\" (UniqueName: \"kubernetes.io/projected/3471f4e0-bf54-427b-9a0b-3b3242ff81e5-kube-api-access-ndbfp\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vjmm\" (UID: \"3471f4e0-bf54-427b-9a0b-3b3242ff81e5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.261620 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.261650 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxdvh\" (UniqueName: \"kubernetes.io/projected/60e9b65b-18c2-45a6-8295-d2dac4a939cd-kube-api-access-kxdvh\") pod \"machine-config-server-j94xz\" (UID: \"60e9b65b-18c2-45a6-8295-d2dac4a939cd\") " pod="openshift-machine-config-operator/machine-config-server-j94xz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.261716 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-serving-cert\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.261744 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/056cfe82-16b2-4c97-bbda-617a10007162-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4pf8r\" (UID: \"056cfe82-16b2-4c97-bbda-617a10007162\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.261900 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4aed05ac-1200-48f7-b840-dc71374cbb51-serving-cert\") pod \"service-ca-operator-777779d784-9bj6c\" (UID: \"4aed05ac-1200-48f7-b840-dc71374cbb51\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.262096 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzkfr\" (UniqueName: \"kubernetes.io/projected/5f63c6f4-9a0c-494e-966d-ccd469dc56c3-kube-api-access-gzkfr\") pod \"olm-operator-6b444d44fb-l7htm\" (UID: \"5f63c6f4-9a0c-494e-966d-ccd469dc56c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.262229 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-csi-data-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.262365 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bj4fz\" (UID: \"fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.262440 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/72170a8d-7210-4a60-9ffc-9871a545fe36-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.262480 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a33960ce-7433-4ae4-a59f-df7a21d886d6-stats-auth\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.262507 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk7xw\" (UniqueName: \"kubernetes.io/projected/4aed05ac-1200-48f7-b840-dc71374cbb51-kube-api-access-zk7xw\") pod \"service-ca-operator-777779d784-9bj6c\" (UID: \"4aed05ac-1200-48f7-b840-dc71374cbb51\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.262567 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42rb5\" (UniqueName: \"kubernetes.io/projected/a134b88a-8fa0-486b-b66d-b1df520f4a6c-kube-api-access-42rb5\") pod \"control-plane-machine-set-operator-78cbb6b69f-fkvxp\" (UID: \"a134b88a-8fa0-486b-b66d-b1df520f4a6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.262654 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-secret-volume\") pod \"collect-profiles-29427435-rlznv\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.262746 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/60e9b65b-18c2-45a6-8295-d2dac4a939cd-node-bootstrap-token\") pod \"machine-config-server-j94xz\" (UID: \"60e9b65b-18c2-45a6-8295-d2dac4a939cd\") " pod="openshift-machine-config-operator/machine-config-server-j94xz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.262825 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnshm\" (UniqueName: \"kubernetes.io/projected/1882ff5e-90e8-456f-a4c8-e8de104b0f46-kube-api-access-vnshm\") pod \"ingress-canary-l68cn\" (UID: \"1882ff5e-90e8-456f-a4c8-e8de104b0f46\") " pod="openshift-ingress-canary/ingress-canary-l68cn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.262877 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a33960ce-7433-4ae4-a59f-df7a21d886d6-metrics-certs\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.264528 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bj4fz\" (UID: \"fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.264989 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-socket-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.265083 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwwh7\" (UniqueName: \"kubernetes.io/projected/af7c90c2-09e5-4f9d-8fa2-81dce6863688-kube-api-access-qwwh7\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:53 crc kubenswrapper[4989]: E1213 17:18:53.265165 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.765140045 +0000 UTC m=+148.371587183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.265306 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cbfd74ca-ade6-4943-95f6-6f930b2d571b-metrics-tls\") pod \"dns-default-ghjq4\" (UID: \"cbfd74ca-ade6-4943-95f6-6f930b2d571b\") " pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.265375 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7abd827d-5201-4f95-95eb-5a817ff7da69-apiservice-cert\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.266468 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dp7s\" (UniqueName: \"kubernetes.io/projected/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-kube-api-access-2dp7s\") pod \"collect-profiles-29427435-rlznv\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.266567 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3471f4e0-bf54-427b-9a0b-3b3242ff81e5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vjmm\" (UID: \"3471f4e0-bf54-427b-9a0b-3b3242ff81e5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.266620 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-config-volume\") pod \"collect-profiles-29427435-rlznv\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.266656 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htgjv\" (UniqueName: \"kubernetes.io/projected/96f26d01-772f-4e9c-b85b-d0585b936f8c-kube-api-access-htgjv\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.266735 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0-srv-cert\") pod \"catalog-operator-68c6474976-7ssm5\" (UID: \"70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.266760 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cbfd74ca-ade6-4943-95f6-6f930b2d571b-config-volume\") pod \"dns-default-ghjq4\" (UID: \"cbfd74ca-ade6-4943-95f6-6f930b2d571b\") " pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.267387 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bj4fz\" (UID: \"fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.267436 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn"] Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.267678 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d96m\" (UniqueName: \"kubernetes.io/projected/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-kube-api-access-5d96m\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.267927 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv9dc\" (UniqueName: \"kubernetes.io/projected/7ee8c55e-83e9-4786-a978-60457031890f-kube-api-access-dv9dc\") pod \"service-ca-9c57cc56f-kklpq\" (UID: \"7ee8c55e-83e9-4786-a978-60457031890f\") " pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.267976 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1882ff5e-90e8-456f-a4c8-e8de104b0f46-cert\") pod \"ingress-canary-l68cn\" (UID: \"1882ff5e-90e8-456f-a4c8-e8de104b0f46\") " pod="openshift-ingress-canary/ingress-canary-l68cn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.268002 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-plugins-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.268294 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgqzp\" (UniqueName: \"kubernetes.io/projected/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-kube-api-access-dgqzp\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.269254 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f960de2c-fee1-4571-a4b4-5f47d796bd88-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9wfgs\" (UID: \"f960de2c-fee1-4571-a4b4-5f47d796bd88\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.270093 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a33960ce-7433-4ae4-a59f-df7a21d886d6-stats-auth\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.271020 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-serving-cert\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.273138 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bj4fz\" (UID: \"fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.273356 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgqzp\" (UniqueName: \"kubernetes.io/projected/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-kube-api-access-dgqzp\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.277741 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htgjv\" (UniqueName: \"kubernetes.io/projected/96f26d01-772f-4e9c-b85b-d0585b936f8c-kube-api-access-htgjv\") pod \"machine-approver-56656f9798-c7m2r\" (UID: \"96f26d01-772f-4e9c-b85b-d0585b936f8c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.282857 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwwh7\" (UniqueName: \"kubernetes.io/projected/af7c90c2-09e5-4f9d-8fa2-81dce6863688-kube-api-access-qwwh7\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.285709 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh64p\" (UniqueName: \"kubernetes.io/projected/72170a8d-7210-4a60-9ffc-9871a545fe36-kube-api-access-nh64p\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.304680 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67vrt\" (UniqueName: \"kubernetes.io/projected/10180d3c-e16a-46d0-a585-63f24d8587cd-kube-api-access-67vrt\") pod \"dns-operator-744455d44c-sq6bg\" (UID: \"10180d3c-e16a-46d0-a585-63f24d8587cd\") " pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.317423 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.332755 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndbfp\" (UniqueName: \"kubernetes.io/projected/3471f4e0-bf54-427b-9a0b-3b3242ff81e5-kube-api-access-ndbfp\") pod \"kube-storage-version-migrator-operator-b67b599dd-9vjmm\" (UID: \"3471f4e0-bf54-427b-9a0b-3b3242ff81e5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.347057 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/72170a8d-7210-4a60-9ffc-9871a545fe36-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9tq69\" (UID: \"72170a8d-7210-4a60-9ffc-9871a545fe36\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.363855 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bppw\" (UniqueName: \"kubernetes.io/projected/a33960ce-7433-4ae4-a59f-df7a21d886d6-kube-api-access-2bppw\") pod \"router-default-5444994796-5t575\" (UID: \"a33960ce-7433-4ae4-a59f-df7a21d886d6\") " pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370345 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370653 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk7xw\" (UniqueName: \"kubernetes.io/projected/4aed05ac-1200-48f7-b840-dc71374cbb51-kube-api-access-zk7xw\") pod \"service-ca-operator-777779d784-9bj6c\" (UID: \"4aed05ac-1200-48f7-b840-dc71374cbb51\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370708 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42rb5\" (UniqueName: \"kubernetes.io/projected/a134b88a-8fa0-486b-b66d-b1df520f4a6c-kube-api-access-42rb5\") pod \"control-plane-machine-set-operator-78cbb6b69f-fkvxp\" (UID: \"a134b88a-8fa0-486b-b66d-b1df520f4a6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370752 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-secret-volume\") pod \"collect-profiles-29427435-rlznv\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370823 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/60e9b65b-18c2-45a6-8295-d2dac4a939cd-node-bootstrap-token\") pod \"machine-config-server-j94xz\" (UID: \"60e9b65b-18c2-45a6-8295-d2dac4a939cd\") " pod="openshift-machine-config-operator/machine-config-server-j94xz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370842 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnshm\" (UniqueName: \"kubernetes.io/projected/1882ff5e-90e8-456f-a4c8-e8de104b0f46-kube-api-access-vnshm\") pod \"ingress-canary-l68cn\" (UID: \"1882ff5e-90e8-456f-a4c8-e8de104b0f46\") " pod="openshift-ingress-canary/ingress-canary-l68cn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370863 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-socket-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370882 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cbfd74ca-ade6-4943-95f6-6f930b2d571b-metrics-tls\") pod \"dns-default-ghjq4\" (UID: \"cbfd74ca-ade6-4943-95f6-6f930b2d571b\") " pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370905 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7abd827d-5201-4f95-95eb-5a817ff7da69-apiservice-cert\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370937 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dp7s\" (UniqueName: \"kubernetes.io/projected/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-kube-api-access-2dp7s\") pod \"collect-profiles-29427435-rlznv\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370975 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-config-volume\") pod \"collect-profiles-29427435-rlznv\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.370994 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0-srv-cert\") pod \"catalog-operator-68c6474976-7ssm5\" (UID: \"70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371015 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cbfd74ca-ade6-4943-95f6-6f930b2d571b-config-volume\") pod \"dns-default-ghjq4\" (UID: \"cbfd74ca-ade6-4943-95f6-6f930b2d571b\") " pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371052 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv9dc\" (UniqueName: \"kubernetes.io/projected/7ee8c55e-83e9-4786-a978-60457031890f-kube-api-access-dv9dc\") pod \"service-ca-9c57cc56f-kklpq\" (UID: \"7ee8c55e-83e9-4786-a978-60457031890f\") " pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371329 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-plugins-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371358 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1882ff5e-90e8-456f-a4c8-e8de104b0f46-cert\") pod \"ingress-canary-l68cn\" (UID: \"1882ff5e-90e8-456f-a4c8-e8de104b0f46\") " pod="openshift-ingress-canary/ingress-canary-l68cn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371381 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f960de2c-fee1-4571-a4b4-5f47d796bd88-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9wfgs\" (UID: \"f960de2c-fee1-4571-a4b4-5f47d796bd88\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371409 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4aed05ac-1200-48f7-b840-dc71374cbb51-config\") pod \"service-ca-operator-777779d784-9bj6c\" (UID: \"4aed05ac-1200-48f7-b840-dc71374cbb51\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371437 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4hmv\" (UniqueName: \"kubernetes.io/projected/cbfd74ca-ade6-4943-95f6-6f930b2d571b-kube-api-access-x4hmv\") pod \"dns-default-ghjq4\" (UID: \"cbfd74ca-ade6-4943-95f6-6f930b2d571b\") " pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371458 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7ee8c55e-83e9-4786-a978-60457031890f-signing-cabundle\") pod \"service-ca-9c57cc56f-kklpq\" (UID: \"7ee8c55e-83e9-4786-a978-60457031890f\") " pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371479 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmr4p\" (UniqueName: \"kubernetes.io/projected/70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0-kube-api-access-rmr4p\") pod \"catalog-operator-68c6474976-7ssm5\" (UID: \"70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371515 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-registration-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371547 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6h6c\" (UniqueName: \"kubernetes.io/projected/7fb1e955-3966-463a-9196-0f9b8ce0f834-kube-api-access-z6h6c\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371569 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-mountpoint-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371594 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r72x6\" (UniqueName: \"kubernetes.io/projected/0607f893-8d28-4498-8ac2-84776eaa26f2-kube-api-access-r72x6\") pod \"migrator-59844c95c7-qbf8q\" (UID: \"0607f893-8d28-4498-8ac2-84776eaa26f2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371615 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/60e9b65b-18c2-45a6-8295-d2dac4a939cd-certs\") pod \"machine-config-server-j94xz\" (UID: \"60e9b65b-18c2-45a6-8295-d2dac4a939cd\") " pod="openshift-machine-config-operator/machine-config-server-j94xz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371652 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7ee8c55e-83e9-4786-a978-60457031890f-signing-key\") pod \"service-ca-9c57cc56f-kklpq\" (UID: \"7ee8c55e-83e9-4786-a978-60457031890f\") " pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371671 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4b2g\" (UniqueName: \"kubernetes.io/projected/056cfe82-16b2-4c97-bbda-617a10007162-kube-api-access-h4b2g\") pod \"package-server-manager-789f6589d5-4pf8r\" (UID: \"056cfe82-16b2-4c97-bbda-617a10007162\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371689 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5f63c6f4-9a0c-494e-966d-ccd469dc56c3-srv-cert\") pod \"olm-operator-6b444d44fb-l7htm\" (UID: \"5f63c6f4-9a0c-494e-966d-ccd469dc56c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371708 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7abd827d-5201-4f95-95eb-5a817ff7da69-tmpfs\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371730 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz9hv\" (UniqueName: \"kubernetes.io/projected/7abd827d-5201-4f95-95eb-5a817ff7da69-kube-api-access-gz9hv\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371756 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzg48\" (UniqueName: \"kubernetes.io/projected/f960de2c-fee1-4571-a4b4-5f47d796bd88-kube-api-access-fzg48\") pod \"multus-admission-controller-857f4d67dd-9wfgs\" (UID: \"f960de2c-fee1-4571-a4b4-5f47d796bd88\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371774 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5f63c6f4-9a0c-494e-966d-ccd469dc56c3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-l7htm\" (UID: \"5f63c6f4-9a0c-494e-966d-ccd469dc56c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371832 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0-profile-collector-cert\") pod \"catalog-operator-68c6474976-7ssm5\" (UID: \"70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371857 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a134b88a-8fa0-486b-b66d-b1df520f4a6c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fkvxp\" (UID: \"a134b88a-8fa0-486b-b66d-b1df520f4a6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371876 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7abd827d-5201-4f95-95eb-5a817ff7da69-webhook-cert\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371917 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxdvh\" (UniqueName: \"kubernetes.io/projected/60e9b65b-18c2-45a6-8295-d2dac4a939cd-kube-api-access-kxdvh\") pod \"machine-config-server-j94xz\" (UID: \"60e9b65b-18c2-45a6-8295-d2dac4a939cd\") " pod="openshift-machine-config-operator/machine-config-server-j94xz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371938 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/056cfe82-16b2-4c97-bbda-617a10007162-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4pf8r\" (UID: \"056cfe82-16b2-4c97-bbda-617a10007162\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371958 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4aed05ac-1200-48f7-b840-dc71374cbb51-serving-cert\") pod \"service-ca-operator-777779d784-9bj6c\" (UID: \"4aed05ac-1200-48f7-b840-dc71374cbb51\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.371992 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzkfr\" (UniqueName: \"kubernetes.io/projected/5f63c6f4-9a0c-494e-966d-ccd469dc56c3-kube-api-access-gzkfr\") pod \"olm-operator-6b444d44fb-l7htm\" (UID: \"5f63c6f4-9a0c-494e-966d-ccd469dc56c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.372010 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-csi-data-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.372146 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-csi-data-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: E1213 17:18:53.372657 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.872640441 +0000 UTC m=+148.479087579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.373090 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-socket-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.373242 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-config-volume\") pod \"collect-profiles-29427435-rlznv\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.377094 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7abd827d-5201-4f95-95eb-5a817ff7da69-apiservice-cert\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.377725 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/60e9b65b-18c2-45a6-8295-d2dac4a939cd-node-bootstrap-token\") pod \"machine-config-server-j94xz\" (UID: \"60e9b65b-18c2-45a6-8295-d2dac4a939cd\") " pod="openshift-machine-config-operator/machine-config-server-j94xz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.378631 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7ee8c55e-83e9-4786-a978-60457031890f-signing-key\") pod \"service-ca-9c57cc56f-kklpq\" (UID: \"7ee8c55e-83e9-4786-a978-60457031890f\") " pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.379584 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/60e9b65b-18c2-45a6-8295-d2dac4a939cd-certs\") pod \"machine-config-server-j94xz\" (UID: \"60e9b65b-18c2-45a6-8295-d2dac4a939cd\") " pod="openshift-machine-config-operator/machine-config-server-j94xz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.381230 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7abd827d-5201-4f95-95eb-5a817ff7da69-tmpfs\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.381255 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-plugins-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.382325 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cbfd74ca-ade6-4943-95f6-6f930b2d571b-config-volume\") pod \"dns-default-ghjq4\" (UID: \"cbfd74ca-ade6-4943-95f6-6f930b2d571b\") " pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.383503 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4aed05ac-1200-48f7-b840-dc71374cbb51-config\") pod \"service-ca-operator-777779d784-9bj6c\" (UID: \"4aed05ac-1200-48f7-b840-dc71374cbb51\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.383552 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a134b88a-8fa0-486b-b66d-b1df520f4a6c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fkvxp\" (UID: \"a134b88a-8fa0-486b-b66d-b1df520f4a6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.383679 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-registration-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.383724 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7fb1e955-3966-463a-9196-0f9b8ce0f834-mountpoint-dir\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.384435 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7ee8c55e-83e9-4786-a978-60457031890f-signing-cabundle\") pod \"service-ca-9c57cc56f-kklpq\" (UID: \"7ee8c55e-83e9-4786-a978-60457031890f\") " pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.384983 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5f63c6f4-9a0c-494e-966d-ccd469dc56c3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-l7htm\" (UID: \"5f63c6f4-9a0c-494e-966d-ccd469dc56c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.385279 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0-profile-collector-cert\") pod \"catalog-operator-68c6474976-7ssm5\" (UID: \"70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.385589 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0-srv-cert\") pod \"catalog-operator-68c6474976-7ssm5\" (UID: \"70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.385707 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5f63c6f4-9a0c-494e-966d-ccd469dc56c3-srv-cert\") pod \"olm-operator-6b444d44fb-l7htm\" (UID: \"5f63c6f4-9a0c-494e-966d-ccd469dc56c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.386458 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4aed05ac-1200-48f7-b840-dc71374cbb51-serving-cert\") pod \"service-ca-operator-777779d784-9bj6c\" (UID: \"4aed05ac-1200-48f7-b840-dc71374cbb51\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.386511 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f960de2c-fee1-4571-a4b4-5f47d796bd88-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9wfgs\" (UID: \"f960de2c-fee1-4571-a4b4-5f47d796bd88\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.386542 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7abd827d-5201-4f95-95eb-5a817ff7da69-webhook-cert\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.387397 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/056cfe82-16b2-4c97-bbda-617a10007162-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4pf8r\" (UID: \"056cfe82-16b2-4c97-bbda-617a10007162\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.388051 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1882ff5e-90e8-456f-a4c8-e8de104b0f46-cert\") pod \"ingress-canary-l68cn\" (UID: \"1882ff5e-90e8-456f-a4c8-e8de104b0f46\") " pod="openshift-ingress-canary/ingress-canary-l68cn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.389250 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.389522 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cbfd74ca-ade6-4943-95f6-6f930b2d571b-metrics-tls\") pod \"dns-default-ghjq4\" (UID: \"cbfd74ca-ade6-4943-95f6-6f930b2d571b\") " pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.389624 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-secret-volume\") pod \"collect-profiles-29427435-rlznv\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.403184 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bj4fz\" (UID: \"fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.423342 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d96m\" (UniqueName: \"kubernetes.io/projected/5bc11da6-8791-4a9f-8ba5-2755cd1c5808-kube-api-access-5d96m\") pod \"console-operator-58897d9998-drznw\" (UID: \"5bc11da6-8791-4a9f-8ba5-2755cd1c5808\") " pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.470015 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnshm\" (UniqueName: \"kubernetes.io/projected/1882ff5e-90e8-456f-a4c8-e8de104b0f46-kube-api-access-vnshm\") pod \"ingress-canary-l68cn\" (UID: \"1882ff5e-90e8-456f-a4c8-e8de104b0f46\") " pod="openshift-ingress-canary/ingress-canary-l68cn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.473975 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474010 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474036 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474085 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474128 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474161 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-config\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474206 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474232 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-images\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474253 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474277 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474303 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474331 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f505d5a-020b-4a22-ae54-62bdfa3985f6-serving-cert\") pod \"openshift-config-operator-7777fb866f-cxb7t\" (UID: \"4f505d5a-020b-4a22-ae54-62bdfa3985f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474365 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-94ggs\" (UID: \"b871b2b1-d492-4db5-91d8-3d77bcdd56e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474381 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.474403 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3be4123f-680d-4afa-a9a4-2a77100953c1-serving-cert\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.477504 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.478255 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.483994 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-l68cn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.484950 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.485514 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b871b2b1-d492-4db5-91d8-3d77bcdd56e9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-94ggs\" (UID: \"b871b2b1-d492-4db5-91d8-3d77bcdd56e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.486956 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:53 crc kubenswrapper[4989]: E1213 17:18:53.487409 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:53.987394733 +0000 UTC m=+148.593841871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.487435 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3be4123f-680d-4afa-a9a4-2a77100953c1-config\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.487916 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.488082 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7fe4787f-cdd3-44fe-8405-dd6ebc9b455c-images\") pod \"machine-api-operator-5694c8668f-8hwgn\" (UID: \"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.488079 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.489593 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dp7s\" (UniqueName: \"kubernetes.io/projected/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-kube-api-access-2dp7s\") pod \"collect-profiles-29427435-rlznv\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.489935 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.490298 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3be4123f-680d-4afa-a9a4-2a77100953c1-serving-cert\") pod \"authentication-operator-69f744f599-7sgxw\" (UID: \"3be4123f-680d-4afa-a9a4-2a77100953c1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.490977 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc182b3d-822b-44aa-a1ba-0af7c2ba5557-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kf2q8\" (UID: \"fc182b3d-822b-44aa-a1ba-0af7c2ba5557\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.494546 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5sdlr\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.495278 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert\") pod \"route-controller-manager-6576b87f9c-4wz97\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.495942 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f505d5a-020b-4a22-ae54-62bdfa3985f6-serving-cert\") pod \"openshift-config-operator-7777fb866f-cxb7t\" (UID: \"4f505d5a-020b-4a22-ae54-62bdfa3985f6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.504006 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk7xw\" (UniqueName: \"kubernetes.io/projected/4aed05ac-1200-48f7-b840-dc71374cbb51-kube-api-access-zk7xw\") pod \"service-ca-operator-777779d784-9bj6c\" (UID: \"4aed05ac-1200-48f7-b840-dc71374cbb51\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.530722 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.540737 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4b2g\" (UniqueName: \"kubernetes.io/projected/056cfe82-16b2-4c97-bbda-617a10007162-kube-api-access-h4b2g\") pod \"package-server-manager-789f6589d5-4pf8r\" (UID: \"056cfe82-16b2-4c97-bbda-617a10007162\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.551629 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz9hv\" (UniqueName: \"kubernetes.io/projected/7abd827d-5201-4f95-95eb-5a817ff7da69-kube-api-access-gz9hv\") pod \"packageserver-d55dfcdfc-7zzhb\" (UID: \"7abd827d-5201-4f95-95eb-5a817ff7da69\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.572145 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv9dc\" (UniqueName: \"kubernetes.io/projected/7ee8c55e-83e9-4786-a978-60457031890f-kube-api-access-dv9dc\") pod \"service-ca-9c57cc56f-kklpq\" (UID: \"7ee8c55e-83e9-4786-a978-60457031890f\") " pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.576225 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:53 crc kubenswrapper[4989]: E1213 17:18:53.576569 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:54.076553392 +0000 UTC m=+148.683000530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.587392 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzg48\" (UniqueName: \"kubernetes.io/projected/f960de2c-fee1-4571-a4b4-5f47d796bd88-kube-api-access-fzg48\") pod \"multus-admission-controller-857f4d67dd-9wfgs\" (UID: \"f960de2c-fee1-4571-a4b4-5f47d796bd88\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.610613 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.613725 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxdvh\" (UniqueName: \"kubernetes.io/projected/60e9b65b-18c2-45a6-8295-d2dac4a939cd-kube-api-access-kxdvh\") pod \"machine-config-server-j94xz\" (UID: \"60e9b65b-18c2-45a6-8295-d2dac4a939cd\") " pod="openshift-machine-config-operator/machine-config-server-j94xz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.631707 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.636306 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42rb5\" (UniqueName: \"kubernetes.io/projected/a134b88a-8fa0-486b-b66d-b1df520f4a6c-kube-api-access-42rb5\") pod \"control-plane-machine-set-operator-78cbb6b69f-fkvxp\" (UID: \"a134b88a-8fa0-486b-b66d-b1df520f4a6c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.646729 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.651848 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmr4p\" (UniqueName: \"kubernetes.io/projected/70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0-kube-api-access-rmr4p\") pod \"catalog-operator-68c6474976-7ssm5\" (UID: \"70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.655903 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-sq6bg"] Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.668642 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.677521 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: E1213 17:18:53.677921 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:54.177907863 +0000 UTC m=+148.784355001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.680151 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzkfr\" (UniqueName: \"kubernetes.io/projected/5f63c6f4-9a0c-494e-966d-ccd469dc56c3-kube-api-access-gzkfr\") pod \"olm-operator-6b444d44fb-l7htm\" (UID: \"5f63c6f4-9a0c-494e-966d-ccd469dc56c3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.690004 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.693350 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4hmv\" (UniqueName: \"kubernetes.io/projected/cbfd74ca-ade6-4943-95f6-6f930b2d571b-kube-api-access-x4hmv\") pod \"dns-default-ghjq4\" (UID: \"cbfd74ca-ade6-4943-95f6-6f930b2d571b\") " pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.711851 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6h6c\" (UniqueName: \"kubernetes.io/projected/7fb1e955-3966-463a-9196-0f9b8ce0f834-kube-api-access-z6h6c\") pod \"csi-hostpathplugin-xqhbq\" (UID: \"7fb1e955-3966-463a-9196-0f9b8ce0f834\") " pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.716710 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.724855 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.729413 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r72x6\" (UniqueName: \"kubernetes.io/projected/0607f893-8d28-4498-8ac2-84776eaa26f2-kube-api-access-r72x6\") pod \"migrator-59844c95c7-qbf8q\" (UID: \"0607f893-8d28-4498-8ac2-84776eaa26f2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.729590 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.731660 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.735986 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.742728 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.749690 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.751123 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.758140 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.760299 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.762886 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.764013 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.771597 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.779853 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.780223 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:53 crc kubenswrapper[4989]: E1213 17:18:53.780503 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:54.280488185 +0000 UTC m=+148.886935323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.804243 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.816070 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.824054 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-j94xz" Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.846071 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-l68cn"] Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.884588 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:53 crc kubenswrapper[4989]: E1213 17:18:53.884942 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:54.384928864 +0000 UTC m=+148.991376002 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:53 crc kubenswrapper[4989]: I1213 17:18:53.985975 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:54 crc kubenswrapper[4989]: E1213 17:18:53.986679 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:54.486625086 +0000 UTC m=+149.093072224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.039403 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q" Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.099817 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:54 crc kubenswrapper[4989]: E1213 17:18:54.100409 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:54.600394628 +0000 UTC m=+149.206841766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.107321 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm"] Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.113931 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9kdlf" event={"ID":"70f58b07-ad87-4ea5-8373-4caae0199fdc","Type":"ContainerStarted","Data":"be30687635f88f8b9211956f6a4d069b62b7f4917fd189f8e51db5da9cfcefcb"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.114288 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9kdlf" event={"ID":"70f58b07-ad87-4ea5-8373-4caae0199fdc","Type":"ContainerStarted","Data":"f15dad3247eb712e185d4364b1369d195fa84b82a317f89ce4ad742a411c71d3"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.117628 4989 generic.go:334] "Generic (PLEG): container finished" podID="18785359-0f0d-4177-a4d2-068af37b0730" containerID="619072729fc952a09e9a4777b43d0a4b3539f576b98a0b7e229d7761f447996a" exitCode=0 Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.129137 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" event={"ID":"18785359-0f0d-4177-a4d2-068af37b0730","Type":"ContainerDied","Data":"619072729fc952a09e9a4777b43d0a4b3539f576b98a0b7e229d7761f447996a"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.129176 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" event={"ID":"18785359-0f0d-4177-a4d2-068af37b0730","Type":"ContainerStarted","Data":"fc5c42f0b7624aac545faa537b995ad0a725b2aefd635b9c08a59d1454bb2128"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.148567 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-drznw"] Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.156293 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-l68cn" event={"ID":"1882ff5e-90e8-456f-a4c8-e8de104b0f46","Type":"ContainerStarted","Data":"af49db5472c4788c8d8374bbe50d48e8c7927114bf502f49312118a9bc3eed66"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.157336 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" event={"ID":"10180d3c-e16a-46d0-a585-63f24d8587cd","Type":"ContainerStarted","Data":"9ed22f9d84e3f3ecf3997ae8ec8750a3bf31d41eca590a834f9de51465d1d28c"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.159409 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" event={"ID":"badbe877-b1f8-40cf-9e7a-1d380a08f2bf","Type":"ContainerStarted","Data":"eeb229e7987e4ce1ac7bba37d265bacd5a8a9258a9033352e5fca9e9798c3e19"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.159533 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" event={"ID":"badbe877-b1f8-40cf-9e7a-1d380a08f2bf","Type":"ContainerStarted","Data":"c0d1984974d3afdd74517ec1ebcfc303dae284feff3e7ff2a4a35bc78bff2296"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.162819 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.167766 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8hwgn"] Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.173063 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69"] Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.183154 4989 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-dvtmv container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.183491 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" podUID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.197934 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" event={"ID":"1b4f9f0c-2b37-4dbe-9dff-ae118b994802","Type":"ContainerStarted","Data":"f924ef21b523795103afb73b72b5a0ae6d2fd9bed30874a273dab6c776c6081d"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.208916 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:54 crc kubenswrapper[4989]: E1213 17:18:54.210505 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:54.710486884 +0000 UTC m=+149.316934022 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.242776 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c03d47c35d636ee8ddf03028c44781ac5ce52b102a3f5acb8f28743435282c53"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.242832 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ce00f3d7b87465a19e0544429302f10eed6c9e879fdda0a5d4c8ed1348b7a09c"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.250295 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-lvjbs" event={"ID":"7b6a2091-2674-4874-a3aa-4b389b8be892","Type":"ContainerStarted","Data":"d8c94c757d67330e62fa951f7bfeb1faf3716c611efd40c7fa2744ce33ef124e"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.250864 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t"] Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.251475 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-lvjbs" Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.258482 4989 generic.go:334] "Generic (PLEG): container finished" podID="3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3" containerID="5cef452c41958cc6ef07d89156b2f4717bb51c05bc1230cc0a773616f9562e33" exitCode=0 Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.258566 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" event={"ID":"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3","Type":"ContainerDied","Data":"5cef452c41958cc6ef07d89156b2f4717bb51c05bc1230cc0a773616f9562e33"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.258619 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" event={"ID":"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3","Type":"ContainerStarted","Data":"f4b4dd34e1786a9b33af14050980629ab05d1d65d103b54dc1321e21cbf684c4"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.268345 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" event={"ID":"96f26d01-772f-4e9c-b85b-d0585b936f8c","Type":"ContainerStarted","Data":"72190fbeb643e7ef7f030ddde0d8c1dc2de55d5b0b699024857ea98aacf5c609"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.270252 4989 patch_prober.go:28] interesting pod/downloads-7954f5f757-lvjbs container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.270313 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lvjbs" podUID="7b6a2091-2674-4874-a3aa-4b389b8be892" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.34:8080/\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.277612 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5sdlr"] Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.279595 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" event={"ID":"06291205-adcc-4352-9c00-ecd417c31f2a","Type":"ContainerStarted","Data":"84953cb8cd2d48ec01387e0e01dbf7338dcc5e250cc4cdb6040c6916455f9d00"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.312536 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:54 crc kubenswrapper[4989]: E1213 17:18:54.314005 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:54.813992313 +0000 UTC m=+149.420439451 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.334523 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" event={"ID":"5cf82975-26bf-461d-af29-79560e79663a","Type":"ContainerStarted","Data":"0c2c167f2c3562a8d1bc74b450e0837d2f061f3843c4a4a26927b3fc6de27c94"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.334588 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" event={"ID":"5cf82975-26bf-461d-af29-79560e79663a","Type":"ContainerStarted","Data":"3e54b4579c94203e2a46d1f6f4cf8208009e9819b79f4379198b6970bf5df89d"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.338209 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"4148d50d1d5b33fbca3b480517f13574371b3910b7c5f3d72d673bfcb2c22bca"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.338244 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5a912c79562e18a6caf708623ff3490030deccb4bcdba47b16beb3788e6a2522"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.349934 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" event={"ID":"bdf2d561-45dc-4672-929e-df9803b1d581","Type":"ContainerStarted","Data":"10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.349979 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" event={"ID":"bdf2d561-45dc-4672-929e-df9803b1d581","Type":"ContainerStarted","Data":"6666f9a0b3c2f434ce57129a75e4b5e616a86120c0b23216c93e6c06c7bcb422"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.350667 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.354562 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5t575" event={"ID":"a33960ce-7433-4ae4-a59f-df7a21d886d6","Type":"ContainerStarted","Data":"3837bde70fe91f7fe9e83314415bc94c9edf1e36a333552a96b0d5c8945812ec"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.360108 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv"] Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.367104 4989 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k5w28 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.367188 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" podUID="bdf2d561-45dc-4672-929e-df9803b1d581" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.398266 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"619d31574a6135a7352db0525fd8ee71c40f016c3e3a820f8716b7ea8caf929e"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.401403 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"49066fafeb22adadee45414264bbf0d67999cdbf94ff31ecb7dceb90efd2239f"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.401466 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.414378 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:54 crc kubenswrapper[4989]: E1213 17:18:54.415953 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:54.915926834 +0000 UTC m=+149.522373972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.492899 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" event={"ID":"78a3473b-2599-43d9-98d5-83f2356618f5","Type":"ContainerStarted","Data":"983085026d5df736427bb01a58ade63b98858f3086968ede4224e7c1315f4b9f"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.492972 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" event={"ID":"78a3473b-2599-43d9-98d5-83f2356618f5","Type":"ContainerStarted","Data":"76d7c3b7184ecb7a9205245e40af831044af72f7df5772067f44f61f2048a942"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.521456 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:54 crc kubenswrapper[4989]: E1213 17:18:54.523329 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:55.023312545 +0000 UTC m=+149.629759683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.565166 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" event={"ID":"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c","Type":"ContainerStarted","Data":"d9d7981997e7e2183ad2a24515b74a848e4a45ecaec7b1287a856ddd8c0d2c9b"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.565237 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" event={"ID":"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c","Type":"ContainerStarted","Data":"bb0aa641aae15e06205d1b0d4addb2066b847c0e2194956d6dbe6b5f0c8d3118"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.565254 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" event={"ID":"04b35497-4fb1-4eae-a6d8-128c5f9aaf5c","Type":"ContainerStarted","Data":"8394a18d7cce7d0dda062f88ff4cdf3a773e22f81e942d78020f980f28364da0"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.606361 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pqk7b" podStartSLOduration=123.606338072 podStartE2EDuration="2m3.606338072s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:54.602393608 +0000 UTC m=+149.208840746" watchObservedRunningTime="2025-12-13 17:18:54.606338072 +0000 UTC m=+149.212785210" Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.641304 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" event={"ID":"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e","Type":"ContainerStarted","Data":"2f691f42204418d5b30ff414b9f8a8b02b29a079e65bb4a53af858c2b608f955"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.641374 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" event={"ID":"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e","Type":"ContainerStarted","Data":"51e217477117e3c27dd324742d405c8a00302a0d95a027567c9ef2ac0a3762c2"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.641386 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" event={"ID":"c71f0f5d-4e9b-4a20-9007-22ec7b5e172e","Type":"ContainerStarted","Data":"eda8a9750050b44305321ecbaa0de90f56fd3a31420dd985f096c21b75544c17"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.652728 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:54 crc kubenswrapper[4989]: E1213 17:18:54.653276 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:55.153239214 +0000 UTC m=+149.759686352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.653575 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:54 crc kubenswrapper[4989]: E1213 17:18:54.654660 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:55.154637038 +0000 UTC m=+149.761084176 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.658766 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7sgxw"] Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.757990 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:54 crc kubenswrapper[4989]: E1213 17:18:54.761029 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:55.261004688 +0000 UTC m=+149.867451826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.811616 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" event={"ID":"d665e361-bdd1-4a87-a913-d67c2d15a102","Type":"ContainerStarted","Data":"195438036f9236fc270b1fbc3e07f44e2d8236cd58592518911368f83fcb1740"} Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.898205 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz"] Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.898359 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:54 crc kubenswrapper[4989]: E1213 17:18:54.898664 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:55.398649828 +0000 UTC m=+150.005096966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.922712 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kklpq"] Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.962300 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-656kn" podStartSLOduration=123.962276836 podStartE2EDuration="2m3.962276836s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:54.920156474 +0000 UTC m=+149.526603602" watchObservedRunningTime="2025-12-13 17:18:54.962276836 +0000 UTC m=+149.568723974" Dec 13 17:18:54 crc kubenswrapper[4989]: I1213 17:18:54.999745 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:55 crc kubenswrapper[4989]: E1213 17:18:55.003461 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:55.503440198 +0000 UTC m=+150.109887336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.102533 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:55 crc kubenswrapper[4989]: E1213 17:18:55.102962 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:55.602949622 +0000 UTC m=+150.209396760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.203435 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v97j6" podStartSLOduration=124.203399526 podStartE2EDuration="2m4.203399526s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:55.18824315 +0000 UTC m=+149.794690288" watchObservedRunningTime="2025-12-13 17:18:55.203399526 +0000 UTC m=+149.809846664" Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.213908 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:55 crc kubenswrapper[4989]: E1213 17:18:55.214688 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:55.71465821 +0000 UTC m=+150.321105348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.310886 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8"] Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.318556 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:55 crc kubenswrapper[4989]: E1213 17:18:55.319197 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:55.819177551 +0000 UTC m=+150.425624689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.342294 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5"] Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.351711 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm"] Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.419477 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:55 crc kubenswrapper[4989]: E1213 17:18:55.419865 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:55.919847031 +0000 UTC m=+150.526294169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.451070 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-9kdlf" podStartSLOduration=124.451052561 podStartE2EDuration="2m4.451052561s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:55.448594314 +0000 UTC m=+150.055041462" watchObservedRunningTime="2025-12-13 17:18:55.451052561 +0000 UTC m=+150.057499689" Dec 13 17:18:55 crc kubenswrapper[4989]: W1213 17:18:55.463089 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc182b3d_822b_44aa_a1ba_0af7c2ba5557.slice/crio-14ab654be9d68259d86f7b24a27e4e6f5f08b281d2090f43daf1fb702f3bc53b WatchSource:0}: Error finding container 14ab654be9d68259d86f7b24a27e4e6f5f08b281d2090f43daf1fb702f3bc53b: Status 404 returned error can't find the container with id 14ab654be9d68259d86f7b24a27e4e6f5f08b281d2090f43daf1fb702f3bc53b Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.524978 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:55 crc kubenswrapper[4989]: E1213 17:18:55.525677 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:56.025663543 +0000 UTC m=+150.632110681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.557810 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7x4jv" podStartSLOduration=124.557779832 podStartE2EDuration="2m4.557779832s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:55.497522169 +0000 UTC m=+150.103969307" watchObservedRunningTime="2025-12-13 17:18:55.557779832 +0000 UTC m=+150.164226960" Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.621702 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-lvjbs" podStartSLOduration=124.621669027 podStartE2EDuration="2m4.621669027s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:55.559686691 +0000 UTC m=+150.166133829" watchObservedRunningTime="2025-12-13 17:18:55.621669027 +0000 UTC m=+150.228116165" Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.636398 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:55 crc kubenswrapper[4989]: E1213 17:18:55.636667 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:56.136649698 +0000 UTC m=+150.743096836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.694854 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97"] Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.735097 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-ncjdp" podStartSLOduration=124.735076678 podStartE2EDuration="2m4.735076678s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:55.702698861 +0000 UTC m=+150.309145999" watchObservedRunningTime="2025-12-13 17:18:55.735076678 +0000 UTC m=+150.341523816" Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.741299 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:55 crc kubenswrapper[4989]: E1213 17:18:55.741591 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:56.241579052 +0000 UTC m=+150.848026180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.754738 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" podStartSLOduration=124.754661373 podStartE2EDuration="2m4.754661373s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:55.732213368 +0000 UTC m=+150.338660506" watchObservedRunningTime="2025-12-13 17:18:55.754661373 +0000 UTC m=+150.361108511" Dec 13 17:18:55 crc kubenswrapper[4989]: I1213 17:18:55.812944 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r"] Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.816562 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c"] Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.842965 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:55.843538 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:56.343513732 +0000 UTC m=+150.949960870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.851302 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp"] Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.858212 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" podStartSLOduration=125.858189713 podStartE2EDuration="2m5.858189713s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:55.839616609 +0000 UTC m=+150.446063747" watchObservedRunningTime="2025-12-13 17:18:55.858189713 +0000 UTC m=+150.464636851" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.897306 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qhk6m" podStartSLOduration=124.89728096 podStartE2EDuration="2m4.89728096s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:55.857849892 +0000 UTC m=+150.464297030" watchObservedRunningTime="2025-12-13 17:18:55.89728096 +0000 UTC m=+150.503728098" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.918972 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9wfgs"] Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.919122 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5t575" event={"ID":"a33960ce-7433-4ae4-a59f-df7a21d886d6","Type":"ContainerStarted","Data":"67413b8c6abc141de4a3fba39958444bac3bbe2dd5dbc95b2e557b94bf2b4bec"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.923081 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q"] Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.923571 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rmmll" podStartSLOduration=124.923547494 podStartE2EDuration="2m4.923547494s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:55.901394359 +0000 UTC m=+150.507841497" watchObservedRunningTime="2025-12-13 17:18:55.923547494 +0000 UTC m=+150.529994632" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.930479 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xqhbq"] Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.940267 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb"] Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.963344 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:55.965104 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:56.465086258 +0000 UTC m=+151.071533386 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.966288 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs"] Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.966979 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-5t575" podStartSLOduration=124.966943557 podStartE2EDuration="2m4.966943557s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:55.962753995 +0000 UTC m=+150.569201123" watchObservedRunningTime="2025-12-13 17:18:55.966943557 +0000 UTC m=+150.573390695" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.972525 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" event={"ID":"96f26d01-772f-4e9c-b85b-d0585b936f8c","Type":"ContainerStarted","Data":"13865f79a74aa511fbb651b44bdb143a9607fc2ea022b66e7628705c858bf2ff"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.974120 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" event={"ID":"fc182b3d-822b-44aa-a1ba-0af7c2ba5557","Type":"ContainerStarted","Data":"14ab654be9d68259d86f7b24a27e4e6f5f08b281d2090f43daf1fb702f3bc53b"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:55.981710 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" event={"ID":"547bcf39-6ef3-4ebd-986b-94fd0b5efaea","Type":"ContainerStarted","Data":"23804f7af592ceda1ed335feafa0d2a1a8e88f74957c15b46e05ec45ce22e721"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.001997 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" event={"ID":"4f505d5a-020b-4a22-ae54-62bdfa3985f6","Type":"ContainerStarted","Data":"5735b1fee31b6b1fcbb6dfdc4bab2b4201037969571d7bf22d25bfc3fecf474a"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.051071 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" event={"ID":"7ee8c55e-83e9-4786-a978-60457031890f","Type":"ContainerStarted","Data":"95a9cba02f20b33ccdf1cf1efb81bdad4e45d39ffaa5ef6fedc1352cfafcfad0"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.051106 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" event={"ID":"af7c90c2-09e5-4f9d-8fa2-81dce6863688","Type":"ContainerStarted","Data":"c3cfbed74a1a14e2fd34a79e2ceda721153875b327cab425d8463c216c4c3513"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.051117 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" event={"ID":"3be4123f-680d-4afa-a9a4-2a77100953c1","Type":"ContainerStarted","Data":"fffe6d017dc8853f610ddbf6357dd7b77ec6a59f5138775bad5e251199895142"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.051127 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" event={"ID":"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c","Type":"ContainerStarted","Data":"55e0a32c8df85b79c81f98168a3fe17d9577f9d5d51bb4cf5c035d9e9f095983"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.051137 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-l68cn" event={"ID":"1882ff5e-90e8-456f-a4c8-e8de104b0f46","Type":"ContainerStarted","Data":"7f46e1000ea2b9bdd45a920cda11de0a9ac263cb8194ca4155cf7c225f9fd67b"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.053562 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" event={"ID":"72170a8d-7210-4a60-9ffc-9871a545fe36","Type":"ContainerStarted","Data":"c602780b0229582fa201d415c2213d87c3b3532cf46b2af608eb4cd309beaf6d"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.062186 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" event={"ID":"3471f4e0-bf54-427b-9a0b-3b3242ff81e5","Type":"ContainerStarted","Data":"01aa21ef4a23697175fed9834a3c1db95cef356f91262ca059ea5a583e3041ac"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.062235 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" event={"ID":"3471f4e0-bf54-427b-9a0b-3b3242ff81e5","Type":"ContainerStarted","Data":"96f556d37371ec2565fca7984bc0d7c74be8a2d9c93a6e5fc6cc84f6df1340fc"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.063895 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:56.064217 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:56.564202911 +0000 UTC m=+151.170650049 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:56 crc kubenswrapper[4989]: W1213 17:18:56.072201 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aed05ac_1200_48f7_b840_dc71374cbb51.slice/crio-b2c160e819ef37bc6da5b2d1dfc289c37b9627451ef6e5698346caf1878a9b97 WatchSource:0}: Error finding container b2c160e819ef37bc6da5b2d1dfc289c37b9627451ef6e5698346caf1878a9b97: Status 404 returned error can't find the container with id b2c160e819ef37bc6da5b2d1dfc289c37b9627451ef6e5698346caf1878a9b97 Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.084023 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-drznw" event={"ID":"5bc11da6-8791-4a9f-8ba5-2755cd1c5808","Type":"ContainerStarted","Data":"6cf7824f3c74cc6f822c655900f580f01b218af464991f0cd22d3f849d05887c"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.084068 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-drznw" event={"ID":"5bc11da6-8791-4a9f-8ba5-2755cd1c5808","Type":"ContainerStarted","Data":"0e2f83a4d3bf505f067fa1e3b5c18cdb4226d12a45aba108dd4f3389a24e10fe"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.090062 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.090108 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-ghjq4"] Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.095654 4989 patch_prober.go:28] interesting pod/console-operator-58897d9998-drznw container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.095691 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-drznw" podUID="5bc11da6-8791-4a9f-8ba5-2755cd1c5808" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.100702 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" event={"ID":"5f63c6f4-9a0c-494e-966d-ccd469dc56c3","Type":"ContainerStarted","Data":"5191017fa559c9616c5dd642fa2dce923db68199f8b26b718dd23e3e5c95d6aa"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.114738 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" event={"ID":"70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0","Type":"ContainerStarted","Data":"5c3c5dc0d51b43c1d1cc5c5d56bf07a19cfbc48785352e6e8a7fc931ff951db6"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.115990 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-j94xz" event={"ID":"60e9b65b-18c2-45a6-8295-d2dac4a939cd","Type":"ContainerStarted","Data":"9160fc4bc0f1e241c62077d9dff1849a216b3c374abf69cd1d18a3dcd287a297"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.117922 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" event={"ID":"fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8","Type":"ContainerStarted","Data":"a5b5cf4981d9de2c6f837c9abbd8ccde4832d315454faed7ead6b127b3b9926d"} Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.130495 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.132427 4989 patch_prober.go:28] interesting pod/downloads-7954f5f757-lvjbs container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.132489 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lvjbs" podUID="7b6a2091-2674-4874-a3aa-4b389b8be892" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.34:8080/\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.168610 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:56.172484 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:56.672470679 +0000 UTC m=+151.278917817 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.192612 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.272682 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:56.277834 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:56.777805867 +0000 UTC m=+151.384253005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.278161 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:56.278610 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:56.778599362 +0000 UTC m=+151.385046500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.382525 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:56.383267 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:56.883250227 +0000 UTC m=+151.489697365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.485096 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:56.485499 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:56.985481317 +0000 UTC m=+151.591928455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.586605 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:56.587282 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:57.087263802 +0000 UTC m=+151.693710940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.613364 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-l68cn" podStartSLOduration=7.613339061 podStartE2EDuration="7.613339061s" podCreationTimestamp="2025-12-13 17:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:56.602063367 +0000 UTC m=+151.208510505" watchObservedRunningTime="2025-12-13 17:18:56.613339061 +0000 UTC m=+151.219786209" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.634242 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.640180 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:18:56 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:18:56 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:18:56 crc kubenswrapper[4989]: healthz check failed Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.640438 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.652703 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-drznw" podStartSLOduration=126.652684006 podStartE2EDuration="2m6.652684006s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:56.651204409 +0000 UTC m=+151.257651547" watchObservedRunningTime="2025-12-13 17:18:56.652684006 +0000 UTC m=+151.259131144" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.689760 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:56.690309 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:57.190291316 +0000 UTC m=+151.796738454 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.734263 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9vjmm" podStartSLOduration=125.734246737 podStartE2EDuration="2m5.734246737s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:56.688763959 +0000 UTC m=+151.295211097" watchObservedRunningTime="2025-12-13 17:18:56.734246737 +0000 UTC m=+151.340693865" Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.792559 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:56.793968 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:57.2939344 +0000 UTC m=+151.900381538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:56 crc kubenswrapper[4989]: I1213 17:18:56.899924 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:56 crc kubenswrapper[4989]: E1213 17:18:56.900288 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:57.400259868 +0000 UTC m=+152.006707006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.001415 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:57 crc kubenswrapper[4989]: E1213 17:18:57.001935 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:57.50191624 +0000 UTC m=+152.108363378 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.104057 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:57 crc kubenswrapper[4989]: E1213 17:18:57.104747 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:57.604732707 +0000 UTC m=+152.211179845 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.141856 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp" event={"ID":"a134b88a-8fa0-486b-b66d-b1df520f4a6c","Type":"ContainerStarted","Data":"b26491a50b016757ac95a74728b87468c1fa5a947781322a3dea8effd2cfbde3"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.143170 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" event={"ID":"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c","Type":"ContainerStarted","Data":"946562364d79d858ff1bb209f37bcb56f1bb8eb8380c2b929099bfc4d7365695"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.153921 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" event={"ID":"3be4123f-680d-4afa-a9a4-2a77100953c1","Type":"ContainerStarted","Data":"c4f1220af9b123af76c01d83e9ed7af7b15c44b5c7b27a18e05687618d55cc82"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.174723 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" event={"ID":"3d6b8fbf-b2c9-4ab7-a269-a3af6cf278e3","Type":"ContainerStarted","Data":"3fb5afe70b33753f30c27bcd17b5a2b19c9d838f8cb94d44db5ba51d397b6d19"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.187616 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-7sgxw" podStartSLOduration=127.187597949 podStartE2EDuration="2m7.187597949s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.187038282 +0000 UTC m=+151.793485430" watchObservedRunningTime="2025-12-13 17:18:57.187597949 +0000 UTC m=+151.794045087" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.190429 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" event={"ID":"f960de2c-fee1-4571-a4b4-5f47d796bd88","Type":"ContainerStarted","Data":"72dfca8d384e5c8a2687c5dcc62a13c4b6c588be3a9158e96369bd9f302a44a2"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.206517 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:57 crc kubenswrapper[4989]: E1213 17:18:57.207828 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:57.707811524 +0000 UTC m=+152.314258662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.212469 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" event={"ID":"10180d3c-e16a-46d0-a585-63f24d8587cd","Type":"ContainerStarted","Data":"a34c0ef057bf1389b3b1cfd0ba3fb6df2c573729760217aef1f217d42e0ecee6"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.214302 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" event={"ID":"fb6d7262-d2b1-4b74-9c27-1bd93ac8a2a8","Type":"ContainerStarted","Data":"0b55deb4e8d25fdd24684a4ebe43f8191a148e6b522202ffcaed7869b6fffe89"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.217455 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" event={"ID":"b871b2b1-d492-4db5-91d8-3d77bcdd56e9","Type":"ContainerStarted","Data":"c37fd48f85fbfdfda30399d1312bd65ace89d160f69de668b65a916884956637"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.222840 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" podStartSLOduration=126.222824715 podStartE2EDuration="2m6.222824715s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.221654859 +0000 UTC m=+151.828101997" watchObservedRunningTime="2025-12-13 17:18:57.222824715 +0000 UTC m=+151.829271853" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.223491 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" event={"ID":"18785359-0f0d-4177-a4d2-068af37b0730","Type":"ContainerStarted","Data":"56b04ecec91f1dfdf3d1f02844ed56bbee7e0fc5c5cdbc9704d81aee0c4954e8"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.243503 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" event={"ID":"72170a8d-7210-4a60-9ffc-9871a545fe36","Type":"ContainerStarted","Data":"be01b69e2eb29f1e4c7b61408da2646f113a8cd399baac32edd2d0856ed089bb"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.256249 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bj4fz" podStartSLOduration=126.256229794 podStartE2EDuration="2m6.256229794s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.253328863 +0000 UTC m=+151.859776001" watchObservedRunningTime="2025-12-13 17:18:57.256229794 +0000 UTC m=+151.862676932" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.264772 4989 generic.go:334] "Generic (PLEG): container finished" podID="4f505d5a-020b-4a22-ae54-62bdfa3985f6" containerID="95df1c1aa2964f1c06d129eca042f26185db724491d4a54023e57fb139cc0591" exitCode=0 Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.266500 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" event={"ID":"4f505d5a-020b-4a22-ae54-62bdfa3985f6","Type":"ContainerDied","Data":"95df1c1aa2964f1c06d129eca042f26185db724491d4a54023e57fb139cc0591"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.269202 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" event={"ID":"af7c90c2-09e5-4f9d-8fa2-81dce6863688","Type":"ContainerStarted","Data":"062c83a59f5d544c965e57eef861a1addcc344328dfce8b5aa5f58587a6f6a34"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.269594 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.278413 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9tq69" podStartSLOduration=126.278396259 podStartE2EDuration="2m6.278396259s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.276944024 +0000 UTC m=+151.883391162" watchObservedRunningTime="2025-12-13 17:18:57.278396259 +0000 UTC m=+151.884843397" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.293300 4989 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-5sdlr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.293364 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" podUID="af7c90c2-09e5-4f9d-8fa2-81dce6863688" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.304186 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" event={"ID":"03635dfa-6663-49c7-a980-9fda1f1f5a29","Type":"ContainerStarted","Data":"0319c52a71ed8dad4dc4f26d8c9d045a6464dcf6375bc9c58d7eb3b42a94130a"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.305126 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.309125 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:57 crc kubenswrapper[4989]: E1213 17:18:57.310429 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:57.810400625 +0000 UTC m=+152.416847763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.313347 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" event={"ID":"056cfe82-16b2-4c97-bbda-617a10007162","Type":"ContainerStarted","Data":"a206262d332ce6305615101841f171264ced303d916618c93d250f33c2bc2e44"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.313407 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" event={"ID":"056cfe82-16b2-4c97-bbda-617a10007162","Type":"ContainerStarted","Data":"e442d79a180c92506f8ab135c075f4ccccdb5c46a94a0a0dab7d0757cc82ca93"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.313521 4989 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4wz97 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.313571 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" podUID="03635dfa-6663-49c7-a980-9fda1f1f5a29" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.314283 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" podStartSLOduration=126.314264156 podStartE2EDuration="2m6.314264156s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.313828242 +0000 UTC m=+151.920275380" watchObservedRunningTime="2025-12-13 17:18:57.314264156 +0000 UTC m=+151.920711294" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.329538 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" event={"ID":"7ee8c55e-83e9-4786-a978-60457031890f","Type":"ContainerStarted","Data":"6437b142c375379c6373ace77d22434d0e10bef52307451967ce1dac01fdfc07"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.346769 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" event={"ID":"547bcf39-6ef3-4ebd-986b-94fd0b5efaea","Type":"ContainerStarted","Data":"ee3a0d16d4f89038ade85cde8412d71580f519a6ba76eb1ab223dbfb9082c251"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.363910 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" event={"ID":"7fb1e955-3966-463a-9196-0f9b8ce0f834","Type":"ContainerStarted","Data":"50f716ad07a83da5964258681c637704c8872d54d71ab663640e384651652f79"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.375837 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" event={"ID":"70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0","Type":"ContainerStarted","Data":"450dd9e7b46dcaea50acdac717c58a4fa17828848f73e2b803a508c51d18fe36"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.376635 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.398927 4989 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-7ssm5 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.398986 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" podUID="70f288d2-1dec-4ba4-8ac6-52fd8e77c3b0" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.404557 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-j94xz" event={"ID":"60e9b65b-18c2-45a6-8295-d2dac4a939cd","Type":"ContainerStarted","Data":"e7c1c253d4e7962a8a9c12851a02f2afd69c3d935701496831808ee2182e4c94"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.410636 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.411679 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" event={"ID":"96f26d01-772f-4e9c-b85b-d0585b936f8c","Type":"ContainerStarted","Data":"6820f554e929c135c14311b63fb9caf1f3cbec9cbe14cdfac5a8444399f38be2"} Dec 13 17:18:57 crc kubenswrapper[4989]: E1213 17:18:57.415359 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:57.915325058 +0000 UTC m=+152.521772246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.423157 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" podStartSLOduration=126.423137684 podStartE2EDuration="2m6.423137684s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.422324438 +0000 UTC m=+152.028771576" watchObservedRunningTime="2025-12-13 17:18:57.423137684 +0000 UTC m=+152.029584822" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.425231 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ghjq4" event={"ID":"cbfd74ca-ade6-4943-95f6-6f930b2d571b","Type":"ContainerStarted","Data":"fdb2593bca11ce119c8950aaf1571fca462c278f3125de9217b19bc356cb9957"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.427355 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-kklpq" podStartSLOduration=126.427342616 podStartE2EDuration="2m6.427342616s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.395971081 +0000 UTC m=+152.002418209" watchObservedRunningTime="2025-12-13 17:18:57.427342616 +0000 UTC m=+152.033789754" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.433632 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" event={"ID":"7abd827d-5201-4f95-95eb-5a817ff7da69","Type":"ContainerStarted","Data":"2f3f55b6695d968ba757f2f7787603dff62c3cf7343321efb6887541cf7918de"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.434895 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.436509 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q" event={"ID":"0607f893-8d28-4498-8ac2-84776eaa26f2","Type":"ContainerStarted","Data":"b01380f4102251ca54483d1a94f82020b6ce0c4dc70491f031ae6829f2f1e604"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.438819 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" event={"ID":"4aed05ac-1200-48f7-b840-dc71374cbb51","Type":"ContainerStarted","Data":"b2c160e819ef37bc6da5b2d1dfc289c37b9627451ef6e5698346caf1878a9b97"} Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.442733 4989 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7zzhb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.442913 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" podUID="7abd827d-5201-4f95-95eb-5a817ff7da69" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.445907 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" podStartSLOduration=126.445892839 podStartE2EDuration="2m6.445892839s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.445366562 +0000 UTC m=+152.051813700" watchObservedRunningTime="2025-12-13 17:18:57.445892839 +0000 UTC m=+152.052339977" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.467294 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-drznw" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.469376 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" podStartSLOduration=127.469355935 podStartE2EDuration="2m7.469355935s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.466995301 +0000 UTC m=+152.073442439" watchObservedRunningTime="2025-12-13 17:18:57.469355935 +0000 UTC m=+152.075803073" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.512396 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:57 crc kubenswrapper[4989]: E1213 17:18:57.513939 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:58.013923014 +0000 UTC m=+152.620370242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.525225 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c7m2r" podStartSLOduration=127.525208558 podStartE2EDuration="2m7.525208558s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.49693029 +0000 UTC m=+152.103377428" watchObservedRunningTime="2025-12-13 17:18:57.525208558 +0000 UTC m=+152.131655696" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.527598 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-j94xz" podStartSLOduration=8.527591003 podStartE2EDuration="8.527591003s" podCreationTimestamp="2025-12-13 17:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.524420743 +0000 UTC m=+152.130867881" watchObservedRunningTime="2025-12-13 17:18:57.527591003 +0000 UTC m=+152.134038141" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.614103 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:57 crc kubenswrapper[4989]: E1213 17:18:57.615056 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:58.115040999 +0000 UTC m=+152.721488137 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.639548 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:18:57 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:18:57 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:18:57 crc kubenswrapper[4989]: healthz check failed Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.639765 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.643685 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" podStartSLOduration=126.643671748 podStartE2EDuration="2m6.643671748s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.613558792 +0000 UTC m=+152.220005930" watchObservedRunningTime="2025-12-13 17:18:57.643671748 +0000 UTC m=+152.250118886" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.649576 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" podStartSLOduration=126.649558052 podStartE2EDuration="2m6.649558052s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:57.641667275 +0000 UTC m=+152.248114413" watchObservedRunningTime="2025-12-13 17:18:57.649558052 +0000 UTC m=+152.256005190" Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.715975 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:57 crc kubenswrapper[4989]: E1213 17:18:57.716545 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:58.216533025 +0000 UTC m=+152.822980163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.817354 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:57 crc kubenswrapper[4989]: E1213 17:18:57.818290 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:58.318268719 +0000 UTC m=+152.924715857 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:57 crc kubenswrapper[4989]: I1213 17:18:57.919581 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:57 crc kubenswrapper[4989]: E1213 17:18:57.920059 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:58.420042234 +0000 UTC m=+153.026489372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.020530 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:58 crc kubenswrapper[4989]: E1213 17:18:58.021106 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:58.521089916 +0000 UTC m=+153.127537054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.122492 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:58 crc kubenswrapper[4989]: E1213 17:18:58.122880 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:58.622865952 +0000 UTC m=+153.229313090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.223474 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:58 crc kubenswrapper[4989]: E1213 17:18:58.223859 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:58.72378082 +0000 UTC m=+153.330227968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.324774 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:58 crc kubenswrapper[4989]: E1213 17:18:58.325181 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:58.825166992 +0000 UTC m=+153.431614130 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.399027 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.400182 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.403642 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.404687 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.409719 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.426361 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:58 crc kubenswrapper[4989]: E1213 17:18:58.426724 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:58.92670481 +0000 UTC m=+153.533151948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.449166 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" event={"ID":"03635dfa-6663-49c7-a980-9fda1f1f5a29","Type":"ContainerStarted","Data":"81262de59a62639ccc2d4e2da270299b828ca257c15d56aac0b33f67f7b7dcfd"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.450510 4989 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4wz97 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.450563 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" podUID="03635dfa-6663-49c7-a980-9fda1f1f5a29" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.452715 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q" event={"ID":"0607f893-8d28-4498-8ac2-84776eaa26f2","Type":"ContainerStarted","Data":"b0d2d548bd3e628cc7c06e620f3eb649a0c2dd1f68399c959586065b7140688d"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.452759 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q" event={"ID":"0607f893-8d28-4498-8ac2-84776eaa26f2","Type":"ContainerStarted","Data":"9305bd7b5a8e3cd319252a8f766d24fc652222b0a5e0d9d3de6f9c45492ade3d"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.455388 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9bj6c" event={"ID":"4aed05ac-1200-48f7-b840-dc71374cbb51","Type":"ContainerStarted","Data":"c37a3af2bc55bb91528d2e212d4cbb4b6c768f4608f918ad5fe1a537aed98985"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.458503 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" event={"ID":"18785359-0f0d-4177-a4d2-068af37b0730","Type":"ContainerStarted","Data":"0c7e76108d03fd49bef1b5beabd683decf8bc4bd6ea6fec843464dc1c1bdf4eb"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.462913 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" event={"ID":"f960de2c-fee1-4571-a4b4-5f47d796bd88","Type":"ContainerStarted","Data":"d21bd2a6f1d027240ad92e1702ef3bc14555f7ab1b926f73d1a1273b62570e65"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.462956 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" event={"ID":"f960de2c-fee1-4571-a4b4-5f47d796bd88","Type":"ContainerStarted","Data":"932f493a5ab47b27f705a4c53f4970311da1fcfb354f8103dcda3fadffa59cc8"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.466734 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" event={"ID":"10180d3c-e16a-46d0-a585-63f24d8587cd","Type":"ContainerStarted","Data":"66dff63017ff6d4149c1482efb60311fe4e6656699c204f6e0a9aaefc46f1d59"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.468895 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" event={"ID":"7fb1e955-3966-463a-9196-0f9b8ce0f834","Type":"ContainerStarted","Data":"812bd2673d6ac14e2852aa9a4bd138a987337a61729ada21f175372b5ccc4248"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.472317 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" event={"ID":"4f505d5a-020b-4a22-ae54-62bdfa3985f6","Type":"ContainerStarted","Data":"e0aa489a9f5dac1da09a68485c997604526b46b7c073221a09724b0154e0417d"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.472510 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.474746 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp" event={"ID":"a134b88a-8fa0-486b-b66d-b1df520f4a6c","Type":"ContainerStarted","Data":"792ae6edf15d2b9e06c6da12f8b64d8c56eeb88d141ba445b70046f5ace60f87"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.484296 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qbf8q" podStartSLOduration=127.484265848 podStartE2EDuration="2m7.484265848s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:58.481225462 +0000 UTC m=+153.087672620" watchObservedRunningTime="2025-12-13 17:18:58.484265848 +0000 UTC m=+153.090712986" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.487226 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ghjq4" event={"ID":"cbfd74ca-ade6-4943-95f6-6f930b2d571b","Type":"ContainerStarted","Data":"c1aaf6cc8ba72a6d6032a6b560d1a1af4b7dabe1d02b2bd3450bbf817a9c7ce2"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.487280 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-ghjq4" event={"ID":"cbfd74ca-ade6-4943-95f6-6f930b2d571b","Type":"ContainerStarted","Data":"01a80bdfaf09d6f3fa02039745f2f4b3ed0b26fe51330319be51f1e52fa1fad2"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.487995 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-ghjq4" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.495565 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" event={"ID":"b871b2b1-d492-4db5-91d8-3d77bcdd56e9","Type":"ContainerStarted","Data":"1db0be321dcdc352db205f2b7355c3aa3d6a3c8203b36f0c587d34cb7b70e913"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.495626 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" event={"ID":"b871b2b1-d492-4db5-91d8-3d77bcdd56e9","Type":"ContainerStarted","Data":"ac4e8b80868e7176d81a0103c0b8dbc57cb5716420e43b7c56364d44bcd5a90d"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.501486 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" event={"ID":"7fe4787f-cdd3-44fe-8405-dd6ebc9b455c","Type":"ContainerStarted","Data":"e76ff8620f86a0e3f5bdf150aaa1869124c35de6050160e100a7eed277618add"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.511233 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" podStartSLOduration=128.511215984 podStartE2EDuration="2m8.511215984s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:58.509235732 +0000 UTC m=+153.115682880" watchObservedRunningTime="2025-12-13 17:18:58.511215984 +0000 UTC m=+153.117663122" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.514038 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" event={"ID":"7abd827d-5201-4f95-95eb-5a817ff7da69","Type":"ContainerStarted","Data":"843115c2a553a9ade68af72b5a4eb0a3c78dce4040988e9c152d4cce3160af9f"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.514585 4989 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7zzhb container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.514624 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" podUID="7abd827d-5201-4f95-95eb-5a817ff7da69" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.531046 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.531220 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6324a8b9-8ed4-416a-af5d-c4563a7414cf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.531423 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6324a8b9-8ed4-416a-af5d-c4563a7414cf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 17:18:58 crc kubenswrapper[4989]: E1213 17:18:58.533901 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.033886796 +0000 UTC m=+153.640333934 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.541104 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" event={"ID":"5f63c6f4-9a0c-494e-966d-ccd469dc56c3","Type":"ContainerStarted","Data":"fccaeb6c6cc924848c0264a6bbd8f1c521999e6b8bcc9895adf435bf42021945"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.541657 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.552555 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" podStartSLOduration=128.552537851 podStartE2EDuration="2m8.552537851s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:58.548879627 +0000 UTC m=+153.155326785" watchObservedRunningTime="2025-12-13 17:18:58.552537851 +0000 UTC m=+153.158984989" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.554268 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" event={"ID":"fc182b3d-822b-44aa-a1ba-0af7c2ba5557","Type":"ContainerStarted","Data":"70e72170a824aeabd2eaf6a62c5fe6181215b5db2cc0398813ba885c1bf6ef64"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.558942 4989 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-l7htm container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.559000 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" podUID="5f63c6f4-9a0c-494e-966d-ccd469dc56c3" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.563695 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" event={"ID":"056cfe82-16b2-4c97-bbda-617a10007162","Type":"ContainerStarted","Data":"4f47162a4f239b114e28e4d8b8d1d204af12b0cd2a828bf9aae6c0858c6b6b10"} Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.563729 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.586181 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.586230 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7ssm5" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.634263 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.634586 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6324a8b9-8ed4-416a-af5d-c4563a7414cf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.635050 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6324a8b9-8ed4-416a-af5d-c4563a7414cf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 17:18:58 crc kubenswrapper[4989]: E1213 17:18:58.636021 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.135990801 +0000 UTC m=+153.742437939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.636386 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6324a8b9-8ed4-416a-af5d-c4563a7414cf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.645533 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:18:58 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:18:58 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:18:58 crc kubenswrapper[4989]: healthz check failed Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.645600 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.649571 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-sq6bg" podStartSLOduration=127.649556197 podStartE2EDuration="2m7.649556197s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:58.640160002 +0000 UTC m=+153.246607160" watchObservedRunningTime="2025-12-13 17:18:58.649556197 +0000 UTC m=+153.256003335" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.658719 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-9wfgs" podStartSLOduration=127.658692873 podStartE2EDuration="2m7.658692873s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:58.605913737 +0000 UTC m=+153.212360875" watchObservedRunningTime="2025-12-13 17:18:58.658692873 +0000 UTC m=+153.265140011" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.691322 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fkvxp" podStartSLOduration=127.691303518 podStartE2EDuration="2m7.691303518s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:58.690084529 +0000 UTC m=+153.296531687" watchObservedRunningTime="2025-12-13 17:18:58.691303518 +0000 UTC m=+153.297750656" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.731392 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6324a8b9-8ed4-416a-af5d-c4563a7414cf\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.749410 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" podStartSLOduration=127.749394121 podStartE2EDuration="2m7.749394121s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:58.747246844 +0000 UTC m=+153.353693982" watchObservedRunningTime="2025-12-13 17:18:58.749394121 +0000 UTC m=+153.355841259" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.752097 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:58 crc kubenswrapper[4989]: E1213 17:18:58.752381 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.252370195 +0000 UTC m=+153.858817333 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.773052 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-ghjq4" podStartSLOduration=9.773035493 podStartE2EDuration="9.773035493s" podCreationTimestamp="2025-12-13 17:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:58.768177431 +0000 UTC m=+153.374624569" watchObservedRunningTime="2025-12-13 17:18:58.773035493 +0000 UTC m=+153.379482631" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.844642 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-8hwgn" podStartSLOduration=127.844625301 podStartE2EDuration="2m7.844625301s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:58.837034073 +0000 UTC m=+153.443481211" watchObservedRunningTime="2025-12-13 17:18:58.844625301 +0000 UTC m=+153.451072439" Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.853164 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:58 crc kubenswrapper[4989]: E1213 17:18:58.853583 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.353558532 +0000 UTC m=+153.960005670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.853634 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:58 crc kubenswrapper[4989]: E1213 17:18:58.854048 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.354038757 +0000 UTC m=+153.960485895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.954734 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:58 crc kubenswrapper[4989]: E1213 17:18:58.956502 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.456486093 +0000 UTC m=+154.062933231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:58 crc kubenswrapper[4989]: I1213 17:18:58.968612 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94ggs" podStartSLOduration=128.968594573 podStartE2EDuration="2m8.968594573s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:58.916854159 +0000 UTC m=+153.523301297" watchObservedRunningTime="2025-12-13 17:18:58.968594573 +0000 UTC m=+153.575041711" Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.019374 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.031972 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" podStartSLOduration=128.031943292 podStartE2EDuration="2m8.031943292s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:59.028558535 +0000 UTC m=+153.635005673" watchObservedRunningTime="2025-12-13 17:18:59.031943292 +0000 UTC m=+153.638390420" Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.057839 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:59 crc kubenswrapper[4989]: E1213 17:18:59.058372 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.558340051 +0000 UTC m=+154.164787189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.098590 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kf2q8" podStartSLOduration=129.098559743 podStartE2EDuration="2m9.098559743s" podCreationTimestamp="2025-12-13 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:18:59.097087687 +0000 UTC m=+153.703534825" watchObservedRunningTime="2025-12-13 17:18:59.098559743 +0000 UTC m=+153.705006881" Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.159653 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:59 crc kubenswrapper[4989]: E1213 17:18:59.159960 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.65994542 +0000 UTC m=+154.266392558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.260939 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:59 crc kubenswrapper[4989]: E1213 17:18:59.261269 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.761256141 +0000 UTC m=+154.367703279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.362043 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:59 crc kubenswrapper[4989]: E1213 17:18:59.362190 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.862172439 +0000 UTC m=+154.468619577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.362770 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:59 crc kubenswrapper[4989]: E1213 17:18:59.363316 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.863293415 +0000 UTC m=+154.469740543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.464728 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:59 crc kubenswrapper[4989]: E1213 17:18:59.465113 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:18:59.96509529 +0000 UTC m=+154.571542428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.566512 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:59 crc kubenswrapper[4989]: E1213 17:18:59.567029 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:19:00.067014441 +0000 UTC m=+154.673461579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.600354 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" event={"ID":"7fb1e955-3966-463a-9196-0f9b8ce0f834","Type":"ContainerStarted","Data":"6af42692813dcc2b82f93b29808ea54fa51e7b82b857c7fb56e5f647236b2437"} Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.605995 4989 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-l7htm container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.606068 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" podUID="5f63c6f4-9a0c-494e-966d-ccd469dc56c3" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.624756 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.644998 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:18:59 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:18:59 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:18:59 crc kubenswrapper[4989]: healthz check failed Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.645060 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.668230 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:59 crc kubenswrapper[4989]: E1213 17:18:59.670171 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:19:00.170156348 +0000 UTC m=+154.776603486 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.772611 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:59 crc kubenswrapper[4989]: E1213 17:18:59.772961 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:19:00.272947625 +0000 UTC m=+154.879394763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.815873 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.873665 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:18:59 crc kubenswrapper[4989]: E1213 17:18:59.874492 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:19:00.374476513 +0000 UTC m=+154.980923641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:18:59 crc kubenswrapper[4989]: I1213 17:18:59.978489 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:18:59 crc kubenswrapper[4989]: E1213 17:18:59.978881 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:19:00.478868991 +0000 UTC m=+155.085316129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.080020 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:19:00 crc kubenswrapper[4989]: E1213 17:19:00.080411 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:19:00.580394618 +0000 UTC m=+155.186841756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.159083 4989 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.181008 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:19:00 crc kubenswrapper[4989]: E1213 17:19:00.181335 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:19:00.681323356 +0000 UTC m=+155.287770494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.251912 4989 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-13T17:19:00.159112379Z","Handler":null,"Name":""} Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.282727 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:19:00 crc kubenswrapper[4989]: E1213 17:19:00.283010 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-13 17:19:00.782969007 +0000 UTC m=+155.389416145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.283080 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:19:00 crc kubenswrapper[4989]: E1213 17:19:00.283532 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-13 17:19:00.783518535 +0000 UTC m=+155.389965673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-9blxx" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.327223 4989 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.327301 4989 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.383921 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d8k9w"] Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.384205 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.385279 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.387898 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.405464 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d8k9w"] Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.441718 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.489650 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-utilities\") pod \"community-operators-d8k9w\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.489730 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-catalog-content\") pod \"community-operators-d8k9w\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.489769 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8nvm\" (UniqueName: \"kubernetes.io/projected/199e44f9-fad2-4498-8c5b-685a10c97ab8-kube-api-access-w8nvm\") pod \"community-operators-d8k9w\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.489842 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.569644 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tbwsb"] Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.570669 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.575357 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.580954 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.581000 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.583083 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7zzhb" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.590500 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-utilities\") pod \"community-operators-d8k9w\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.591024 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-utilities\") pod \"community-operators-d8k9w\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.594212 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-catalog-content\") pod \"community-operators-d8k9w\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.594303 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8nvm\" (UniqueName: \"kubernetes.io/projected/199e44f9-fad2-4498-8c5b-685a10c97ab8-kube-api-access-w8nvm\") pod \"community-operators-d8k9w\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.594533 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-catalog-content\") pod \"community-operators-d8k9w\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.602634 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tbwsb"] Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.608858 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" event={"ID":"7fb1e955-3966-463a-9196-0f9b8ce0f834","Type":"ContainerStarted","Data":"ba6e417f0090ade6a7dec91afad07de8f5a5ea36d939cd067d0c448dd6a0e0d5"} Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.608898 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" event={"ID":"7fb1e955-3966-463a-9196-0f9b8ce0f834","Type":"ContainerStarted","Data":"6ebf7f0d7ea06c1bdfa7514460c2ad1a2fff028ee8ebb127b88209dd44664bf4"} Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.610430 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6324a8b9-8ed4-416a-af5d-c4563a7414cf","Type":"ContainerStarted","Data":"58d3eacce90c9a47d370067d77f75554cf7286fee0de9010ca56db820852030c"} Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.610456 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6324a8b9-8ed4-416a-af5d-c4563a7414cf","Type":"ContainerStarted","Data":"c0b1a4a3ef76c781ae7d48c1401ca4945654237096e214c41c80306ec8802ac0"} Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.636023 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-l7htm" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.636873 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:00 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:00 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:00 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.636950 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.654358 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8nvm\" (UniqueName: \"kubernetes.io/projected/199e44f9-fad2-4498-8c5b-685a10c97ab8-kube-api-access-w8nvm\") pod \"community-operators-d8k9w\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.694670 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-9blxx\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.695851 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-utilities\") pod \"certified-operators-tbwsb\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.695995 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtv88\" (UniqueName: \"kubernetes.io/projected/5fda1871-6df5-4256-8a90-74bb43a13290-kube-api-access-qtv88\") pod \"certified-operators-tbwsb\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.696096 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-catalog-content\") pod \"certified-operators-tbwsb\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.702463 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.788265 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.798822 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtv88\" (UniqueName: \"kubernetes.io/projected/5fda1871-6df5-4256-8a90-74bb43a13290-kube-api-access-qtv88\") pod \"certified-operators-tbwsb\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.798953 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-catalog-content\") pod \"certified-operators-tbwsb\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.799029 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-utilities\") pod \"certified-operators-tbwsb\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.800108 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-catalog-content\") pod \"certified-operators-tbwsb\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.800141 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-utilities\") pod \"certified-operators-tbwsb\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.861902 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cdfb5"] Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.874418 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:00 crc kubenswrapper[4989]: I1213 17:19:00.893990 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtv88\" (UniqueName: \"kubernetes.io/projected/5fda1871-6df5-4256-8a90-74bb43a13290-kube-api-access-qtv88\") pod \"certified-operators-tbwsb\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.009577 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p644\" (UniqueName: \"kubernetes.io/projected/b9b8349c-c894-46a8-a043-adebd14c0dac-kube-api-access-8p644\") pod \"community-operators-cdfb5\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.009648 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-utilities\") pod \"community-operators-cdfb5\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.009761 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-catalog-content\") pod \"community-operators-cdfb5\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.074224 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-xqhbq" podStartSLOduration=12.074191198 podStartE2EDuration="12.074191198s" podCreationTimestamp="2025-12-13 17:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:19:00.924165097 +0000 UTC m=+155.530612235" watchObservedRunningTime="2025-12-13 17:19:01.074191198 +0000 UTC m=+155.680638336" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.086403 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cdfb5"] Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.089893 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g5fxh"] Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.092918 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.098603 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.098562613 podStartE2EDuration="3.098562613s" podCreationTimestamp="2025-12-13 17:18:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:19:01.062207221 +0000 UTC m=+155.668654359" watchObservedRunningTime="2025-12-13 17:19:01.098562613 +0000 UTC m=+155.705009751" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.111094 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-catalog-content\") pod \"community-operators-cdfb5\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.111163 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p644\" (UniqueName: \"kubernetes.io/projected/b9b8349c-c894-46a8-a043-adebd14c0dac-kube-api-access-8p644\") pod \"community-operators-cdfb5\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.111195 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-utilities\") pod \"community-operators-cdfb5\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.111680 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-utilities\") pod \"community-operators-cdfb5\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.111933 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-catalog-content\") pod \"community-operators-cdfb5\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.194471 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.195413 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g5fxh"] Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.196570 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p644\" (UniqueName: \"kubernetes.io/projected/b9b8349c-c894-46a8-a043-adebd14c0dac-kube-api-access-8p644\") pod \"community-operators-cdfb5\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.214856 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpz9q\" (UniqueName: \"kubernetes.io/projected/7eeee1b2-8ab2-4635-b805-e30362bf7582-kube-api-access-fpz9q\") pod \"certified-operators-g5fxh\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.214986 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-catalog-content\") pod \"certified-operators-g5fxh\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.215048 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-utilities\") pod \"certified-operators-g5fxh\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.316357 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-catalog-content\") pod \"certified-operators-g5fxh\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.316804 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-utilities\") pod \"certified-operators-g5fxh\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.316827 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpz9q\" (UniqueName: \"kubernetes.io/projected/7eeee1b2-8ab2-4635-b805-e30362bf7582-kube-api-access-fpz9q\") pod \"certified-operators-g5fxh\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.317440 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-catalog-content\") pod \"certified-operators-g5fxh\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.317455 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-utilities\") pod \"certified-operators-g5fxh\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.363637 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpz9q\" (UniqueName: \"kubernetes.io/projected/7eeee1b2-8ab2-4635-b805-e30362bf7582-kube-api-access-fpz9q\") pod \"certified-operators-g5fxh\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.422610 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.439081 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d8k9w"] Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.493260 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.640279 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:01 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:01 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:01 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.640343 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.690732 4989 generic.go:334] "Generic (PLEG): container finished" podID="6324a8b9-8ed4-416a-af5d-c4563a7414cf" containerID="58d3eacce90c9a47d370067d77f75554cf7286fee0de9010ca56db820852030c" exitCode=0 Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.690825 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6324a8b9-8ed4-416a-af5d-c4563a7414cf","Type":"ContainerDied","Data":"58d3eacce90c9a47d370067d77f75554cf7286fee0de9010ca56db820852030c"} Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.706929 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8k9w" event={"ID":"199e44f9-fad2-4498-8c5b-685a10c97ab8","Type":"ContainerStarted","Data":"347203dc3f1b8e5d74c7205e16fa93ebb4acb2705f257180850b5881aefed598"} Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.740382 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9blxx"] Dec 13 17:19:01 crc kubenswrapper[4989]: I1213 17:19:01.950280 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tbwsb"] Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.022603 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.043775 4989 patch_prober.go:28] interesting pod/downloads-7954f5f757-lvjbs container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.34:8080/\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.043835 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-lvjbs" podUID="7b6a2091-2674-4874-a3aa-4b389b8be892" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.34:8080/\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.043960 4989 patch_prober.go:28] interesting pod/downloads-7954f5f757-lvjbs container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.044000 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-lvjbs" podUID="7b6a2091-2674-4874-a3aa-4b389b8be892" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.34:8080/\": dial tcp 10.217.0.34:8080: connect: connection refused" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.079858 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.080281 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.100248 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.100883 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.106206 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.106868 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.107122 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.123716 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.135309 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g5fxh"] Dec 13 17:19:02 crc kubenswrapper[4989]: W1213 17:19:02.153976 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7eeee1b2_8ab2_4635_b805_e30362bf7582.slice/crio-89e3bbab87fd6d8e3750f48cddd058d1746db4f5ce25b9f9e99a4244600bd3de WatchSource:0}: Error finding container 89e3bbab87fd6d8e3750f48cddd058d1746db4f5ce25b9f9e99a4244600bd3de: Status 404 returned error can't find the container with id 89e3bbab87fd6d8e3750f48cddd058d1746db4f5ce25b9f9e99a4244600bd3de Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.253946 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cdfb5"] Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.256435 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.256500 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.321409 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.321660 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.339779 4989 patch_prober.go:28] interesting pod/console-f9d7485db-9kdlf container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.339862 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9kdlf" podUID="70f58b07-ad87-4ea5-8373-4caae0199fdc" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.367484 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.367554 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.367940 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.384438 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.384489 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.388111 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.402319 4989 patch_prober.go:28] interesting pod/apiserver-76f77b778f-2c6kj container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]log ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]etcd ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]poststarthook/generic-apiserver-start-informers ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]poststarthook/max-in-flight-filter ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 13 17:19:02 crc kubenswrapper[4989]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 13 17:19:02 crc kubenswrapper[4989]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]poststarthook/project.openshift.io-projectcache ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]poststarthook/openshift.io-startinformers ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 13 17:19:02 crc kubenswrapper[4989]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 13 17:19:02 crc kubenswrapper[4989]: livez check failed Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.402403 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" podUID="18785359-0f0d-4177-a4d2-068af37b0730" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.473877 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.589076 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x6f77"] Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.590497 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.596809 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.622271 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6f77"] Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.637391 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:02 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:02 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:02 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.637478 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.747215 4989 generic.go:334] "Generic (PLEG): container finished" podID="547bcf39-6ef3-4ebd-986b-94fd0b5efaea" containerID="ee3a0d16d4f89038ade85cde8412d71580f519a6ba76eb1ab223dbfb9082c251" exitCode=0 Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.747382 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" event={"ID":"547bcf39-6ef3-4ebd-986b-94fd0b5efaea","Type":"ContainerDied","Data":"ee3a0d16d4f89038ade85cde8412d71580f519a6ba76eb1ab223dbfb9082c251"} Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.756262 4989 generic.go:334] "Generic (PLEG): container finished" podID="7eeee1b2-8ab2-4635-b805-e30362bf7582" containerID="15f119ecc45725dc4e04dd710af40f456be116658724b37554071547e1c8ac2c" exitCode=0 Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.756339 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5fxh" event={"ID":"7eeee1b2-8ab2-4635-b805-e30362bf7582","Type":"ContainerDied","Data":"15f119ecc45725dc4e04dd710af40f456be116658724b37554071547e1c8ac2c"} Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.756373 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5fxh" event={"ID":"7eeee1b2-8ab2-4635-b805-e30362bf7582","Type":"ContainerStarted","Data":"89e3bbab87fd6d8e3750f48cddd058d1746db4f5ce25b9f9e99a4244600bd3de"} Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.763383 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.768625 4989 generic.go:334] "Generic (PLEG): container finished" podID="b9b8349c-c894-46a8-a043-adebd14c0dac" containerID="02d37f727b2895cc05a5584fc1eab665958baee5d544adb930be48c579ed5e75" exitCode=0 Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.768732 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdfb5" event={"ID":"b9b8349c-c894-46a8-a043-adebd14c0dac","Type":"ContainerDied","Data":"02d37f727b2895cc05a5584fc1eab665958baee5d544adb930be48c579ed5e75"} Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.768771 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdfb5" event={"ID":"b9b8349c-c894-46a8-a043-adebd14c0dac","Type":"ContainerStarted","Data":"41176d7be3701f61707831b317c311518fad7c53f766201000ddcf6b9f4ae019"} Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.773314 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-utilities\") pod \"redhat-marketplace-x6f77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.773351 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-catalog-content\") pod \"redhat-marketplace-x6f77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.773379 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmj9m\" (UniqueName: \"kubernetes.io/projected/00527548-b884-4adf-8c65-4c6ca4367d77-kube-api-access-vmj9m\") pod \"redhat-marketplace-x6f77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.776827 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cxb7t" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.782271 4989 generic.go:334] "Generic (PLEG): container finished" podID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerID="cb57bfd12d0420a459b71aa75c063c99bd1b5b6bcb393c579f15a0f8f451d3fc" exitCode=0 Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.782855 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8k9w" event={"ID":"199e44f9-fad2-4498-8c5b-685a10c97ab8","Type":"ContainerDied","Data":"cb57bfd12d0420a459b71aa75c063c99bd1b5b6bcb393c579f15a0f8f451d3fc"} Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.787378 4989 generic.go:334] "Generic (PLEG): container finished" podID="5fda1871-6df5-4256-8a90-74bb43a13290" containerID="8a3da68912b07caabef9849d5882773774bc75c66ae98f10845555017a583045" exitCode=0 Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.787446 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tbwsb" event={"ID":"5fda1871-6df5-4256-8a90-74bb43a13290","Type":"ContainerDied","Data":"8a3da68912b07caabef9849d5882773774bc75c66ae98f10845555017a583045"} Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.787475 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tbwsb" event={"ID":"5fda1871-6df5-4256-8a90-74bb43a13290","Type":"ContainerStarted","Data":"0754b9172aa9d21d604466e8af3d73f4cf7ba11627ec142a5b46e446393c403e"} Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.824930 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" event={"ID":"fca2d650-00f0-4f7e-9735-540fe4d2a0e0","Type":"ContainerStarted","Data":"1c9fe4feadc2a115ba69b9808cc226433034f03a80e262e53cb0e1ff8311fd28"} Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.824990 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" event={"ID":"fca2d650-00f0-4f7e-9735-540fe4d2a0e0","Type":"ContainerStarted","Data":"e28856604f0c54f3a335e61e671ec16e53868fa7ff88e4d4d2a376c704241490"} Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.850081 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-sjxxg" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.898982 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmj9m\" (UniqueName: \"kubernetes.io/projected/00527548-b884-4adf-8c65-4c6ca4367d77-kube-api-access-vmj9m\") pod \"redhat-marketplace-x6f77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.899106 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-utilities\") pod \"redhat-marketplace-x6f77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.899138 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-catalog-content\") pod \"redhat-marketplace-x6f77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.899653 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-catalog-content\") pod \"redhat-marketplace-x6f77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.901087 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-utilities\") pod \"redhat-marketplace-x6f77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.956874 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmj9m\" (UniqueName: \"kubernetes.io/projected/00527548-b884-4adf-8c65-4c6ca4367d77-kube-api-access-vmj9m\") pod \"redhat-marketplace-x6f77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.970477 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5m6vx"] Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.976829 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:19:02 crc kubenswrapper[4989]: I1213 17:19:02.998902 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m6vx"] Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.071018 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" podStartSLOduration=132.070995816 podStartE2EDuration="2m12.070995816s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:19:03.017153526 +0000 UTC m=+157.623600664" watchObservedRunningTime="2025-12-13 17:19:03.070995816 +0000 UTC m=+157.677442954" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.101880 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-utilities\") pod \"redhat-marketplace-5m6vx\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.101969 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xjk8\" (UniqueName: \"kubernetes.io/projected/157d815b-6695-4cc3-b83a-e0816757a4c4-kube-api-access-6xjk8\") pod \"redhat-marketplace-5m6vx\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.102004 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-catalog-content\") pod \"redhat-marketplace-5m6vx\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.121316 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.203056 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xjk8\" (UniqueName: \"kubernetes.io/projected/157d815b-6695-4cc3-b83a-e0816757a4c4-kube-api-access-6xjk8\") pod \"redhat-marketplace-5m6vx\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.203114 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-catalog-content\") pod \"redhat-marketplace-5m6vx\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.203223 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-utilities\") pod \"redhat-marketplace-5m6vx\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.203735 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-utilities\") pod \"redhat-marketplace-5m6vx\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.204321 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-catalog-content\") pod \"redhat-marketplace-5m6vx\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.209760 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.226773 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xjk8\" (UniqueName: \"kubernetes.io/projected/157d815b-6695-4cc3-b83a-e0816757a4c4-kube-api-access-6xjk8\") pod \"redhat-marketplace-5m6vx\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.335988 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.338533 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.509244 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kube-api-access\") pod \"6324a8b9-8ed4-416a-af5d-c4563a7414cf\" (UID: \"6324a8b9-8ed4-416a-af5d-c4563a7414cf\") " Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.509316 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kubelet-dir\") pod \"6324a8b9-8ed4-416a-af5d-c4563a7414cf\" (UID: \"6324a8b9-8ed4-416a-af5d-c4563a7414cf\") " Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.509735 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6324a8b9-8ed4-416a-af5d-c4563a7414cf" (UID: "6324a8b9-8ed4-416a-af5d-c4563a7414cf"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.527171 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6324a8b9-8ed4-416a-af5d-c4563a7414cf" (UID: "6324a8b9-8ed4-416a-af5d-c4563a7414cf"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.564742 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6f77"] Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.574971 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j5xss"] Dec 13 17:19:03 crc kubenswrapper[4989]: E1213 17:19:03.575223 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6324a8b9-8ed4-416a-af5d-c4563a7414cf" containerName="pruner" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.575243 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6324a8b9-8ed4-416a-af5d-c4563a7414cf" containerName="pruner" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.575405 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6324a8b9-8ed4-416a-af5d-c4563a7414cf" containerName="pruner" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.577200 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.583778 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.598258 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j5xss"] Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.611021 4989 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.611063 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6324a8b9-8ed4-416a-af5d-c4563a7414cf-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:03 crc kubenswrapper[4989]: W1213 17:19:03.611968 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00527548_b884_4adf_8c65_4c6ca4367d77.slice/crio-dfa1ac17780b89489726deadd00d46d6b0522c9988333eb88be4dfaee66b39d5 WatchSource:0}: Error finding container dfa1ac17780b89489726deadd00d46d6b0522c9988333eb88be4dfaee66b39d5: Status 404 returned error can't find the container with id dfa1ac17780b89489726deadd00d46d6b0522c9988333eb88be4dfaee66b39d5 Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.632523 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.638721 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:03 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:03 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:03 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.638821 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.712088 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-utilities\") pod \"redhat-operators-j5xss\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.712127 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-catalog-content\") pod \"redhat-operators-j5xss\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.712145 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdrlp\" (UniqueName: \"kubernetes.io/projected/37118126-08b8-46bd-bde4-a4faace357a2-kube-api-access-zdrlp\") pod \"redhat-operators-j5xss\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.756354 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m6vx"] Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.813954 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-utilities\") pod \"redhat-operators-j5xss\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.814002 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-catalog-content\") pod \"redhat-operators-j5xss\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.814022 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdrlp\" (UniqueName: \"kubernetes.io/projected/37118126-08b8-46bd-bde4-a4faace357a2-kube-api-access-zdrlp\") pod \"redhat-operators-j5xss\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.814655 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-utilities\") pod \"redhat-operators-j5xss\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.815196 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-catalog-content\") pod \"redhat-operators-j5xss\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.842387 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdrlp\" (UniqueName: \"kubernetes.io/projected/37118126-08b8-46bd-bde4-a4faace357a2-kube-api-access-zdrlp\") pod \"redhat-operators-j5xss\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.848728 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5","Type":"ContainerStarted","Data":"86676065ef15f038d70ceb5f1e4890752781b45b62a662838ba3c07116818ff3"} Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.850360 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6324a8b9-8ed4-416a-af5d-c4563a7414cf","Type":"ContainerDied","Data":"c0b1a4a3ef76c781ae7d48c1401ca4945654237096e214c41c80306ec8802ac0"} Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.850383 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0b1a4a3ef76c781ae7d48c1401ca4945654237096e214c41c80306ec8802ac0" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.850536 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.851322 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6f77" event={"ID":"00527548-b884-4adf-8c65-4c6ca4367d77","Type":"ContainerStarted","Data":"dfa1ac17780b89489726deadd00d46d6b0522c9988333eb88be4dfaee66b39d5"} Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.853131 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m6vx" event={"ID":"157d815b-6695-4cc3-b83a-e0816757a4c4","Type":"ContainerStarted","Data":"bd460960bab9a2d211bbaac5c9040acf33d384747720dc93eab05f8f6b6b1720"} Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.856555 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.899827 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.962316 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zlhq7"] Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.971254 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:03 crc kubenswrapper[4989]: I1213 17:19:03.982605 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zlhq7"] Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.123337 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-catalog-content\") pod \"redhat-operators-zlhq7\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.123379 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdd6v\" (UniqueName: \"kubernetes.io/projected/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-kube-api-access-fdd6v\") pod \"redhat-operators-zlhq7\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.123413 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-utilities\") pod \"redhat-operators-zlhq7\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.190522 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.214312 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j5xss"] Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.228607 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-catalog-content\") pod \"redhat-operators-zlhq7\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.228670 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdd6v\" (UniqueName: \"kubernetes.io/projected/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-kube-api-access-fdd6v\") pod \"redhat-operators-zlhq7\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.228712 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-utilities\") pod \"redhat-operators-zlhq7\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.229441 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-utilities\") pod \"redhat-operators-zlhq7\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.229726 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-catalog-content\") pod \"redhat-operators-zlhq7\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.252654 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdd6v\" (UniqueName: \"kubernetes.io/projected/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-kube-api-access-fdd6v\") pod \"redhat-operators-zlhq7\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.291747 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.329856 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-secret-volume\") pod \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.329995 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-config-volume\") pod \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.330090 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dp7s\" (UniqueName: \"kubernetes.io/projected/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-kube-api-access-2dp7s\") pod \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\" (UID: \"547bcf39-6ef3-4ebd-986b-94fd0b5efaea\") " Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.331018 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-config-volume" (OuterVolumeSpecName: "config-volume") pod "547bcf39-6ef3-4ebd-986b-94fd0b5efaea" (UID: "547bcf39-6ef3-4ebd-986b-94fd0b5efaea"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.333777 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-kube-api-access-2dp7s" (OuterVolumeSpecName: "kube-api-access-2dp7s") pod "547bcf39-6ef3-4ebd-986b-94fd0b5efaea" (UID: "547bcf39-6ef3-4ebd-986b-94fd0b5efaea"). InnerVolumeSpecName "kube-api-access-2dp7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.334270 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "547bcf39-6ef3-4ebd-986b-94fd0b5efaea" (UID: "547bcf39-6ef3-4ebd-986b-94fd0b5efaea"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.432665 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dp7s\" (UniqueName: \"kubernetes.io/projected/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-kube-api-access-2dp7s\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.432696 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.432705 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/547bcf39-6ef3-4ebd-986b-94fd0b5efaea-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.574416 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zlhq7"] Dec 13 17:19:04 crc kubenswrapper[4989]: W1213 17:19:04.585202 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16260c45_c803_4e92_b71d_dfb9cd5b8ea2.slice/crio-bd1d861b2adfe9d57e327665f36f9a2d68cf4e1fe16a62c1c7e64d60994c5f13 WatchSource:0}: Error finding container bd1d861b2adfe9d57e327665f36f9a2d68cf4e1fe16a62c1c7e64d60994c5f13: Status 404 returned error can't find the container with id bd1d861b2adfe9d57e327665f36f9a2d68cf4e1fe16a62c1c7e64d60994c5f13 Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.649306 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:04 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:04 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:04 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.649903 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.863088 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5xss" event={"ID":"37118126-08b8-46bd-bde4-a4faace357a2","Type":"ContainerStarted","Data":"21d5c11dd6ceed384e4462b7097e56d2b5d4f66c1c756b20c476dc7575fd81ba"} Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.868780 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zlhq7" event={"ID":"16260c45-c803-4e92-b71d-dfb9cd5b8ea2","Type":"ContainerStarted","Data":"bd1d861b2adfe9d57e327665f36f9a2d68cf4e1fe16a62c1c7e64d60994c5f13"} Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.872083 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.872065 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29427435-rlznv" event={"ID":"547bcf39-6ef3-4ebd-986b-94fd0b5efaea","Type":"ContainerDied","Data":"23804f7af592ceda1ed335feafa0d2a1a8e88f74957c15b46e05ec45ce22e721"} Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.872205 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23804f7af592ceda1ed335feafa0d2a1a8e88f74957c15b46e05ec45ce22e721" Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.875560 4989 generic.go:334] "Generic (PLEG): container finished" podID="157d815b-6695-4cc3-b83a-e0816757a4c4" containerID="a7441b88eb35a3f756f16f2d9eaa631a92a5e28c10f391f6f7d487b6a4c56379" exitCode=0 Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.875631 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m6vx" event={"ID":"157d815b-6695-4cc3-b83a-e0816757a4c4","Type":"ContainerDied","Data":"a7441b88eb35a3f756f16f2d9eaa631a92a5e28c10f391f6f7d487b6a4c56379"} Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.877690 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5","Type":"ContainerStarted","Data":"792655f43fb693ddf0d3d4c639fbd186e7a104c433e19711e5f1a2b467b4d21c"} Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.879434 4989 generic.go:334] "Generic (PLEG): container finished" podID="00527548-b884-4adf-8c65-4c6ca4367d77" containerID="df86ff0474ef1e17bd90f9237fd5a38c3c3f55e819bb38d37e2067b5ef0d518e" exitCode=0 Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.879490 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6f77" event={"ID":"00527548-b884-4adf-8c65-4c6ca4367d77","Type":"ContainerDied","Data":"df86ff0474ef1e17bd90f9237fd5a38c3c3f55e819bb38d37e2067b5ef0d518e"} Dec 13 17:19:04 crc kubenswrapper[4989]: I1213 17:19:04.895506 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.8954819450000002 podStartE2EDuration="2.895481945s" podCreationTimestamp="2025-12-13 17:19:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:19:04.890348573 +0000 UTC m=+159.496795731" watchObservedRunningTime="2025-12-13 17:19:04.895481945 +0000 UTC m=+159.501929093" Dec 13 17:19:05 crc kubenswrapper[4989]: I1213 17:19:05.642276 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:05 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:05 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:05 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:05 crc kubenswrapper[4989]: I1213 17:19:05.642347 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:05 crc kubenswrapper[4989]: I1213 17:19:05.888534 4989 generic.go:334] "Generic (PLEG): container finished" podID="37118126-08b8-46bd-bde4-a4faace357a2" containerID="d35400cb1c4a5dbdd7d518e731cffebf5ed10600f07df3e07ebbecd133d3fff9" exitCode=0 Dec 13 17:19:05 crc kubenswrapper[4989]: I1213 17:19:05.888610 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5xss" event={"ID":"37118126-08b8-46bd-bde4-a4faace357a2","Type":"ContainerDied","Data":"d35400cb1c4a5dbdd7d518e731cffebf5ed10600f07df3e07ebbecd133d3fff9"} Dec 13 17:19:05 crc kubenswrapper[4989]: I1213 17:19:05.890726 4989 generic.go:334] "Generic (PLEG): container finished" podID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerID="95d192dd0234f05a0d70de672c617494fcc159e617838b5d8cbd6ed9f463d1c5" exitCode=0 Dec 13 17:19:05 crc kubenswrapper[4989]: I1213 17:19:05.890832 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zlhq7" event={"ID":"16260c45-c803-4e92-b71d-dfb9cd5b8ea2","Type":"ContainerDied","Data":"95d192dd0234f05a0d70de672c617494fcc159e617838b5d8cbd6ed9f463d1c5"} Dec 13 17:19:05 crc kubenswrapper[4989]: I1213 17:19:05.893511 4989 generic.go:334] "Generic (PLEG): container finished" podID="0bc8e4a3-ee8c-4679-860d-3f4c68d740e5" containerID="792655f43fb693ddf0d3d4c639fbd186e7a104c433e19711e5f1a2b467b4d21c" exitCode=0 Dec 13 17:19:05 crc kubenswrapper[4989]: I1213 17:19:05.893592 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5","Type":"ContainerDied","Data":"792655f43fb693ddf0d3d4c639fbd186e7a104c433e19711e5f1a2b467b4d21c"} Dec 13 17:19:06 crc kubenswrapper[4989]: I1213 17:19:06.643573 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:06 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:06 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:06 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:06 crc kubenswrapper[4989]: I1213 17:19:06.643626 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.250155 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.391746 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.397586 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-2c6kj" Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.405365 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kube-api-access\") pod \"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5\" (UID: \"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5\") " Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.405416 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kubelet-dir\") pod \"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5\" (UID: \"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5\") " Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.405743 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0bc8e4a3-ee8c-4679-860d-3f4c68d740e5" (UID: "0bc8e4a3-ee8c-4679-860d-3f4c68d740e5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.446722 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0bc8e4a3-ee8c-4679-860d-3f4c68d740e5" (UID: "0bc8e4a3-ee8c-4679-860d-3f4c68d740e5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.508181 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.508218 4989 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0bc8e4a3-ee8c-4679-860d-3f4c68d740e5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.638655 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:07 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:07 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:07 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.638709 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.910617 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0bc8e4a3-ee8c-4679-860d-3f4c68d740e5","Type":"ContainerDied","Data":"86676065ef15f038d70ceb5f1e4890752781b45b62a662838ba3c07116818ff3"} Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.910667 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 13 17:19:07 crc kubenswrapper[4989]: I1213 17:19:07.910694 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86676065ef15f038d70ceb5f1e4890752781b45b62a662838ba3c07116818ff3" Dec 13 17:19:08 crc kubenswrapper[4989]: I1213 17:19:08.644087 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:08 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:08 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:08 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:08 crc kubenswrapper[4989]: I1213 17:19:08.644163 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:08 crc kubenswrapper[4989]: I1213 17:19:08.822757 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-ghjq4" Dec 13 17:19:08 crc kubenswrapper[4989]: I1213 17:19:08.830624 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:19:08 crc kubenswrapper[4989]: I1213 17:19:08.830685 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:19:09 crc kubenswrapper[4989]: I1213 17:19:09.643063 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:09 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:09 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:09 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:09 crc kubenswrapper[4989]: I1213 17:19:09.643472 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:10 crc kubenswrapper[4989]: I1213 17:19:10.643598 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:10 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:10 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:10 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:10 crc kubenswrapper[4989]: I1213 17:19:10.643663 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:11 crc kubenswrapper[4989]: I1213 17:19:11.635296 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:11 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:11 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:11 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:11 crc kubenswrapper[4989]: I1213 17:19:11.635398 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:12 crc kubenswrapper[4989]: I1213 17:19:12.052212 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-lvjbs" Dec 13 17:19:12 crc kubenswrapper[4989]: I1213 17:19:12.311880 4989 patch_prober.go:28] interesting pod/console-f9d7485db-9kdlf container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 13 17:19:12 crc kubenswrapper[4989]: I1213 17:19:12.311955 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9kdlf" podUID="70f58b07-ad87-4ea5-8373-4caae0199fdc" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 13 17:19:12 crc kubenswrapper[4989]: I1213 17:19:12.640430 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:12 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:12 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:12 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:12 crc kubenswrapper[4989]: I1213 17:19:12.640834 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:13 crc kubenswrapper[4989]: I1213 17:19:13.458618 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:19:13 crc kubenswrapper[4989]: I1213 17:19:13.468262 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d912915-788e-412a-bae8-6eccd6b4c238-metrics-certs\") pod \"network-metrics-daemon-lfpf8\" (UID: \"7d912915-788e-412a-bae8-6eccd6b4c238\") " pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:19:13 crc kubenswrapper[4989]: I1213 17:19:13.591911 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lfpf8" Dec 13 17:19:13 crc kubenswrapper[4989]: I1213 17:19:13.639334 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:13 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Dec 13 17:19:13 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:13 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:13 crc kubenswrapper[4989]: I1213 17:19:13.639424 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:14 crc kubenswrapper[4989]: I1213 17:19:14.634685 4989 patch_prober.go:28] interesting pod/router-default-5444994796-5t575 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 13 17:19:14 crc kubenswrapper[4989]: [+]has-synced ok Dec 13 17:19:14 crc kubenswrapper[4989]: [+]process-running ok Dec 13 17:19:14 crc kubenswrapper[4989]: healthz check failed Dec 13 17:19:14 crc kubenswrapper[4989]: I1213 17:19:14.634874 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5t575" podUID="a33960ce-7433-4ae4-a59f-df7a21d886d6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 13 17:19:15 crc kubenswrapper[4989]: I1213 17:19:15.637531 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:19:15 crc kubenswrapper[4989]: I1213 17:19:15.651096 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-5t575" Dec 13 17:19:16 crc kubenswrapper[4989]: I1213 17:19:16.974907 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5sdlr"] Dec 13 17:19:16 crc kubenswrapper[4989]: I1213 17:19:16.976135 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" podUID="af7c90c2-09e5-4f9d-8fa2-81dce6863688" containerName="controller-manager" containerID="cri-o://062c83a59f5d544c965e57eef861a1addcc344328dfce8b5aa5f58587a6f6a34" gracePeriod=30 Dec 13 17:19:16 crc kubenswrapper[4989]: I1213 17:19:16.986324 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97"] Dec 13 17:19:16 crc kubenswrapper[4989]: I1213 17:19:16.986669 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" podUID="03635dfa-6663-49c7-a980-9fda1f1f5a29" containerName="route-controller-manager" containerID="cri-o://81262de59a62639ccc2d4e2da270299b828ca257c15d56aac0b33f67f7b7dcfd" gracePeriod=30 Dec 13 17:19:18 crc kubenswrapper[4989]: I1213 17:19:18.057414 4989 generic.go:334] "Generic (PLEG): container finished" podID="af7c90c2-09e5-4f9d-8fa2-81dce6863688" containerID="062c83a59f5d544c965e57eef861a1addcc344328dfce8b5aa5f58587a6f6a34" exitCode=0 Dec 13 17:19:18 crc kubenswrapper[4989]: I1213 17:19:18.057516 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" event={"ID":"af7c90c2-09e5-4f9d-8fa2-81dce6863688","Type":"ContainerDied","Data":"062c83a59f5d544c965e57eef861a1addcc344328dfce8b5aa5f58587a6f6a34"} Dec 13 17:19:18 crc kubenswrapper[4989]: I1213 17:19:18.059821 4989 generic.go:334] "Generic (PLEG): container finished" podID="03635dfa-6663-49c7-a980-9fda1f1f5a29" containerID="81262de59a62639ccc2d4e2da270299b828ca257c15d56aac0b33f67f7b7dcfd" exitCode=0 Dec 13 17:19:18 crc kubenswrapper[4989]: I1213 17:19:18.059861 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" event={"ID":"03635dfa-6663-49c7-a980-9fda1f1f5a29","Type":"ContainerDied","Data":"81262de59a62639ccc2d4e2da270299b828ca257c15d56aac0b33f67f7b7dcfd"} Dec 13 17:19:20 crc kubenswrapper[4989]: I1213 17:19:20.793572 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:19:22 crc kubenswrapper[4989]: I1213 17:19:22.317899 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:19:22 crc kubenswrapper[4989]: I1213 17:19:22.330112 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-9kdlf" Dec 13 17:19:23 crc kubenswrapper[4989]: I1213 17:19:23.670031 4989 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-5sdlr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 13 17:19:23 crc kubenswrapper[4989]: I1213 17:19:23.670098 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" podUID="af7c90c2-09e5-4f9d-8fa2-81dce6863688" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 13 17:19:23 crc kubenswrapper[4989]: I1213 17:19:23.977050 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.007264 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt"] Dec 13 17:19:24 crc kubenswrapper[4989]: E1213 17:19:24.007512 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="547bcf39-6ef3-4ebd-986b-94fd0b5efaea" containerName="collect-profiles" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.007528 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="547bcf39-6ef3-4ebd-986b-94fd0b5efaea" containerName="collect-profiles" Dec 13 17:19:24 crc kubenswrapper[4989]: E1213 17:19:24.007547 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc8e4a3-ee8c-4679-860d-3f4c68d740e5" containerName="pruner" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.007555 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc8e4a3-ee8c-4679-860d-3f4c68d740e5" containerName="pruner" Dec 13 17:19:24 crc kubenswrapper[4989]: E1213 17:19:24.007578 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03635dfa-6663-49c7-a980-9fda1f1f5a29" containerName="route-controller-manager" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.007588 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="03635dfa-6663-49c7-a980-9fda1f1f5a29" containerName="route-controller-manager" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.007701 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bc8e4a3-ee8c-4679-860d-3f4c68d740e5" containerName="pruner" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.007720 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="03635dfa-6663-49c7-a980-9fda1f1f5a29" containerName="route-controller-manager" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.007733 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="547bcf39-6ef3-4ebd-986b-94fd0b5efaea" containerName="collect-profiles" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.008286 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.030771 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmx5l\" (UniqueName: \"kubernetes.io/projected/03635dfa-6663-49c7-a980-9fda1f1f5a29-kube-api-access-qmx5l\") pod \"03635dfa-6663-49c7-a980-9fda1f1f5a29\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.031499 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca\") pod \"03635dfa-6663-49c7-a980-9fda1f1f5a29\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.031556 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert\") pod \"03635dfa-6663-49c7-a980-9fda1f1f5a29\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.031612 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config\") pod \"03635dfa-6663-49c7-a980-9fda1f1f5a29\" (UID: \"03635dfa-6663-49c7-a980-9fda1f1f5a29\") " Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.031922 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-serving-cert\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.032028 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-config\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.032075 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-client-ca\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.032311 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l755p\" (UniqueName: \"kubernetes.io/projected/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-kube-api-access-l755p\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.032455 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca" (OuterVolumeSpecName: "client-ca") pod "03635dfa-6663-49c7-a980-9fda1f1f5a29" (UID: "03635dfa-6663-49c7-a980-9fda1f1f5a29"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.034355 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config" (OuterVolumeSpecName: "config") pod "03635dfa-6663-49c7-a980-9fda1f1f5a29" (UID: "03635dfa-6663-49c7-a980-9fda1f1f5a29"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.039226 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt"] Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.039816 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "03635dfa-6663-49c7-a980-9fda1f1f5a29" (UID: "03635dfa-6663-49c7-a980-9fda1f1f5a29"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.053998 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03635dfa-6663-49c7-a980-9fda1f1f5a29-kube-api-access-qmx5l" (OuterVolumeSpecName: "kube-api-access-qmx5l") pod "03635dfa-6663-49c7-a980-9fda1f1f5a29" (UID: "03635dfa-6663-49c7-a980-9fda1f1f5a29"). InnerVolumeSpecName "kube-api-access-qmx5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.096476 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" event={"ID":"03635dfa-6663-49c7-a980-9fda1f1f5a29","Type":"ContainerDied","Data":"0319c52a71ed8dad4dc4f26d8c9d045a6464dcf6375bc9c58d7eb3b42a94130a"} Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.096556 4989 scope.go:117] "RemoveContainer" containerID="81262de59a62639ccc2d4e2da270299b828ca257c15d56aac0b33f67f7b7dcfd" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.096730 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.128903 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97"] Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.131651 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97"] Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.133753 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-client-ca\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.133887 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l755p\" (UniqueName: \"kubernetes.io/projected/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-kube-api-access-l755p\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.133947 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-serving-cert\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.134024 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-config\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.134111 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.134135 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03635dfa-6663-49c7-a980-9fda1f1f5a29-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.134150 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03635dfa-6663-49c7-a980-9fda1f1f5a29-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.134169 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmx5l\" (UniqueName: \"kubernetes.io/projected/03635dfa-6663-49c7-a980-9fda1f1f5a29-kube-api-access-qmx5l\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.135674 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-config\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.136089 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-client-ca\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.140778 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-serving-cert\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.156175 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l755p\" (UniqueName: \"kubernetes.io/projected/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-kube-api-access-l755p\") pod \"route-controller-manager-78f64778cd-qvqwt\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.338873 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.761926 4989 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4wz97 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: i/o timeout (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 13 17:19:24 crc kubenswrapper[4989]: I1213 17:19:24.762261 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4wz97" podUID="03635dfa-6663-49c7-a980-9fda1f1f5a29" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: i/o timeout (Client.Timeout exceeded while awaiting headers)" Dec 13 17:19:26 crc kubenswrapper[4989]: I1213 17:19:26.021235 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03635dfa-6663-49c7-a980-9fda1f1f5a29" path="/var/lib/kubelet/pods/03635dfa-6663-49c7-a980-9fda1f1f5a29/volumes" Dec 13 17:19:30 crc kubenswrapper[4989]: I1213 17:19:30.614444 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lfpf8"] Dec 13 17:19:30 crc kubenswrapper[4989]: E1213 17:19:30.661424 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 13 17:19:30 crc kubenswrapper[4989]: E1213 17:19:30.661591 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8p644,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cdfb5_openshift-marketplace(b9b8349c-c894-46a8-a043-adebd14c0dac): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 17:19:30 crc kubenswrapper[4989]: E1213 17:19:30.664142 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cdfb5" podUID="b9b8349c-c894-46a8-a043-adebd14c0dac" Dec 13 17:19:31 crc kubenswrapper[4989]: E1213 17:19:31.582195 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 13 17:19:31 crc kubenswrapper[4989]: E1213 17:19:31.582346 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w8nvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-d8k9w_openshift-marketplace(199e44f9-fad2-4498-8c5b-685a10c97ab8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 17:19:31 crc kubenswrapper[4989]: E1213 17:19:31.583673 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-d8k9w" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" Dec 13 17:19:32 crc kubenswrapper[4989]: I1213 17:19:32.526596 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 13 17:19:32 crc kubenswrapper[4989]: E1213 17:19:32.962879 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cdfb5" podUID="b9b8349c-c894-46a8-a043-adebd14c0dac" Dec 13 17:19:32 crc kubenswrapper[4989]: E1213 17:19:32.963272 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d8k9w" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" Dec 13 17:19:33 crc kubenswrapper[4989]: E1213 17:19:33.305044 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 13 17:19:33 crc kubenswrapper[4989]: E1213 17:19:33.305235 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fpz9q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-g5fxh_openshift-marketplace(7eeee1b2-8ab2-4635-b805-e30362bf7582): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 17:19:33 crc kubenswrapper[4989]: E1213 17:19:33.306570 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-g5fxh" podUID="7eeee1b2-8ab2-4635-b805-e30362bf7582" Dec 13 17:19:33 crc kubenswrapper[4989]: E1213 17:19:33.395359 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 13 17:19:33 crc kubenswrapper[4989]: E1213 17:19:33.396116 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qtv88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tbwsb_openshift-marketplace(5fda1871-6df5-4256-8a90-74bb43a13290): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 17:19:33 crc kubenswrapper[4989]: E1213 17:19:33.397317 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-tbwsb" podUID="5fda1871-6df5-4256-8a90-74bb43a13290" Dec 13 17:19:33 crc kubenswrapper[4989]: I1213 17:19:33.748828 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4pf8r" Dec 13 17:19:34 crc kubenswrapper[4989]: I1213 17:19:34.669896 4989 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-5sdlr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 13 17:19:34 crc kubenswrapper[4989]: I1213 17:19:34.669981 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" podUID="af7c90c2-09e5-4f9d-8fa2-81dce6863688" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 13 17:19:36 crc kubenswrapper[4989]: E1213 17:19:36.658023 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 13 17:19:36 crc kubenswrapper[4989]: E1213 17:19:36.659330 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zdrlp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-j5xss_openshift-marketplace(37118126-08b8-46bd-bde4-a4faace357a2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 17:19:36 crc kubenswrapper[4989]: E1213 17:19:36.661023 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-j5xss" podUID="37118126-08b8-46bd-bde4-a4faace357a2" Dec 13 17:19:36 crc kubenswrapper[4989]: E1213 17:19:36.675206 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 13 17:19:36 crc kubenswrapper[4989]: E1213 17:19:36.675402 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fdd6v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zlhq7_openshift-marketplace(16260c45-c803-4e92-b71d-dfb9cd5b8ea2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 17:19:36 crc kubenswrapper[4989]: E1213 17:19:36.676585 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zlhq7" podUID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" Dec 13 17:19:36 crc kubenswrapper[4989]: I1213 17:19:36.918539 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt"] Dec 13 17:19:38 crc kubenswrapper[4989]: E1213 17:19:38.019342 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-g5fxh" podUID="7eeee1b2-8ab2-4635-b805-e30362bf7582" Dec 13 17:19:38 crc kubenswrapper[4989]: E1213 17:19:38.019384 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-j5xss" podUID="37118126-08b8-46bd-bde4-a4faace357a2" Dec 13 17:19:38 crc kubenswrapper[4989]: E1213 17:19:38.019457 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tbwsb" podUID="5fda1871-6df5-4256-8a90-74bb43a13290" Dec 13 17:19:38 crc kubenswrapper[4989]: E1213 17:19:38.078268 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 13 17:19:38 crc kubenswrapper[4989]: E1213 17:19:38.078430 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6xjk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5m6vx_openshift-marketplace(157d815b-6695-4cc3-b83a-e0816757a4c4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 17:19:38 crc kubenswrapper[4989]: E1213 17:19:38.079701 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5m6vx" podUID="157d815b-6695-4cc3-b83a-e0816757a4c4" Dec 13 17:19:38 crc kubenswrapper[4989]: E1213 17:19:38.087771 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 13 17:19:38 crc kubenswrapper[4989]: E1213 17:19:38.087950 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vmj9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-x6f77_openshift-marketplace(00527548-b884-4adf-8c65-4c6ca4367d77): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 13 17:19:38 crc kubenswrapper[4989]: E1213 17:19:38.090201 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-x6f77" podUID="00527548-b884-4adf-8c65-4c6ca4367d77" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.112346 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.157242 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-77fbccf6-5bj5c"] Dec 13 17:19:38 crc kubenswrapper[4989]: E1213 17:19:38.158290 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af7c90c2-09e5-4f9d-8fa2-81dce6863688" containerName="controller-manager" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.158308 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="af7c90c2-09e5-4f9d-8fa2-81dce6863688" containerName="controller-manager" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.158550 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="af7c90c2-09e5-4f9d-8fa2-81dce6863688" containerName="controller-manager" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.159339 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.172367 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77fbccf6-5bj5c"] Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.173891 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles\") pod \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.174436 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config\") pod \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.174480 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca\") pod \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.175281 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af7c90c2-09e5-4f9d-8fa2-81dce6863688-serving-cert\") pod \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.175335 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwwh7\" (UniqueName: \"kubernetes.io/projected/af7c90c2-09e5-4f9d-8fa2-81dce6863688-kube-api-access-qwwh7\") pod \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\" (UID: \"af7c90c2-09e5-4f9d-8fa2-81dce6863688\") " Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.175639 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-client-ca\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.175777 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vx7b\" (UniqueName: \"kubernetes.io/projected/e2bdda06-6da0-49de-84e5-f2d34ff27a29-kube-api-access-8vx7b\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.175870 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2bdda06-6da0-49de-84e5-f2d34ff27a29-serving-cert\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.175910 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-proxy-ca-bundles\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.175943 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-config\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.175946 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca" (OuterVolumeSpecName: "client-ca") pod "af7c90c2-09e5-4f9d-8fa2-81dce6863688" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.176072 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "af7c90c2-09e5-4f9d-8fa2-81dce6863688" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.176772 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config" (OuterVolumeSpecName: "config") pod "af7c90c2-09e5-4f9d-8fa2-81dce6863688" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.180507 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" event={"ID":"af7c90c2-09e5-4f9d-8fa2-81dce6863688","Type":"ContainerDied","Data":"c3cfbed74a1a14e2fd34a79e2ceda721153875b327cab425d8463c216c4c3513"} Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.180591 4989 scope.go:117] "RemoveContainer" containerID="062c83a59f5d544c965e57eef861a1addcc344328dfce8b5aa5f58587a6f6a34" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.181481 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5sdlr" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.184140 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af7c90c2-09e5-4f9d-8fa2-81dce6863688-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "af7c90c2-09e5-4f9d-8fa2-81dce6863688" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.191008 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af7c90c2-09e5-4f9d-8fa2-81dce6863688-kube-api-access-qwwh7" (OuterVolumeSpecName: "kube-api-access-qwwh7") pod "af7c90c2-09e5-4f9d-8fa2-81dce6863688" (UID: "af7c90c2-09e5-4f9d-8fa2-81dce6863688"). InnerVolumeSpecName "kube-api-access-qwwh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.209345 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" event={"ID":"7d912915-788e-412a-bae8-6eccd6b4c238","Type":"ContainerStarted","Data":"1b7ea4f2d6932d928dedfb423f88dd0b7187d3b4f96aea87e950cef79016210c"} Dec 13 17:19:38 crc kubenswrapper[4989]: E1213 17:19:38.214112 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-x6f77" podUID="00527548-b884-4adf-8c65-4c6ca4367d77" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.266869 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt"] Dec 13 17:19:38 crc kubenswrapper[4989]: W1213 17:19:38.271988 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1af1e529_7c4b_4ccc_94ac_2143a4ce014e.slice/crio-78528becf881a19400b383d080f770f91dc50d30673732e7d0ec6d14e77135f6 WatchSource:0}: Error finding container 78528becf881a19400b383d080f770f91dc50d30673732e7d0ec6d14e77135f6: Status 404 returned error can't find the container with id 78528becf881a19400b383d080f770f91dc50d30673732e7d0ec6d14e77135f6 Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.277171 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2bdda06-6da0-49de-84e5-f2d34ff27a29-serving-cert\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.277216 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-proxy-ca-bundles\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.277244 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-config\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.277277 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-client-ca\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.277332 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vx7b\" (UniqueName: \"kubernetes.io/projected/e2bdda06-6da0-49de-84e5-f2d34ff27a29-kube-api-access-8vx7b\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.277373 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af7c90c2-09e5-4f9d-8fa2-81dce6863688-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.277394 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwwh7\" (UniqueName: \"kubernetes.io/projected/af7c90c2-09e5-4f9d-8fa2-81dce6863688-kube-api-access-qwwh7\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.277408 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.277418 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.277428 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af7c90c2-09e5-4f9d-8fa2-81dce6863688-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.278723 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-config\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.280518 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-client-ca\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.285480 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2bdda06-6da0-49de-84e5-f2d34ff27a29-serving-cert\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.288696 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-proxy-ca-bundles\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.293428 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vx7b\" (UniqueName: \"kubernetes.io/projected/e2bdda06-6da0-49de-84e5-f2d34ff27a29-kube-api-access-8vx7b\") pod \"controller-manager-77fbccf6-5bj5c\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.495621 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.526154 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5sdlr"] Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.529655 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5sdlr"] Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.829899 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.830300 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:19:38 crc kubenswrapper[4989]: I1213 17:19:38.946269 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77fbccf6-5bj5c"] Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.217914 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" event={"ID":"1af1e529-7c4b-4ccc-94ac-2143a4ce014e","Type":"ContainerStarted","Data":"89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57"} Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.217984 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" event={"ID":"1af1e529-7c4b-4ccc-94ac-2143a4ce014e","Type":"ContainerStarted","Data":"78528becf881a19400b383d080f770f91dc50d30673732e7d0ec6d14e77135f6"} Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.218114 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" podUID="1af1e529-7c4b-4ccc-94ac-2143a4ce014e" containerName="route-controller-manager" containerID="cri-o://89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57" gracePeriod=30 Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.218277 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.223386 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" event={"ID":"7d912915-788e-412a-bae8-6eccd6b4c238","Type":"ContainerStarted","Data":"25e191999a034ee3c9fbb3e545e9f058092345e06fb2d5d37838727ccd3c808d"} Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.223449 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lfpf8" event={"ID":"7d912915-788e-412a-bae8-6eccd6b4c238","Type":"ContainerStarted","Data":"77842aa11acc265afd7c6e9dd4f8d84203ba7d339faca38f6d04f2b79d1f6002"} Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.223811 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.225164 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" event={"ID":"e2bdda06-6da0-49de-84e5-f2d34ff27a29","Type":"ContainerStarted","Data":"2c45facf861b25f5c184ff376131589b69a1d01f52b022ff6d9653f2a7d3bafc"} Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.263325 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-lfpf8" podStartSLOduration=168.263300251 podStartE2EDuration="2m48.263300251s" podCreationTimestamp="2025-12-13 17:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:19:39.253102412 +0000 UTC m=+193.859549560" watchObservedRunningTime="2025-12-13 17:19:39.263300251 +0000 UTC m=+193.869747379" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.264482 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" podStartSLOduration=22.264472779 podStartE2EDuration="22.264472779s" podCreationTimestamp="2025-12-13 17:19:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:19:39.239000539 +0000 UTC m=+193.845447737" watchObservedRunningTime="2025-12-13 17:19:39.264472779 +0000 UTC m=+193.870919917" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.608313 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.703274 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-config\") pod \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.703357 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-serving-cert\") pod \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.703428 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-client-ca\") pod \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.703471 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l755p\" (UniqueName: \"kubernetes.io/projected/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-kube-api-access-l755p\") pod \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\" (UID: \"1af1e529-7c4b-4ccc-94ac-2143a4ce014e\") " Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.704599 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-client-ca" (OuterVolumeSpecName: "client-ca") pod "1af1e529-7c4b-4ccc-94ac-2143a4ce014e" (UID: "1af1e529-7c4b-4ccc-94ac-2143a4ce014e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.705033 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-config" (OuterVolumeSpecName: "config") pod "1af1e529-7c4b-4ccc-94ac-2143a4ce014e" (UID: "1af1e529-7c4b-4ccc-94ac-2143a4ce014e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.711308 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-kube-api-access-l755p" (OuterVolumeSpecName: "kube-api-access-l755p") pod "1af1e529-7c4b-4ccc-94ac-2143a4ce014e" (UID: "1af1e529-7c4b-4ccc-94ac-2143a4ce014e"). InnerVolumeSpecName "kube-api-access-l755p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.713980 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1af1e529-7c4b-4ccc-94ac-2143a4ce014e" (UID: "1af1e529-7c4b-4ccc-94ac-2143a4ce014e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.804932 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.804989 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l755p\" (UniqueName: \"kubernetes.io/projected/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-kube-api-access-l755p\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.805001 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:39 crc kubenswrapper[4989]: I1213 17:19:39.805012 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1af1e529-7c4b-4ccc-94ac-2143a4ce014e-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.020452 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af7c90c2-09e5-4f9d-8fa2-81dce6863688" path="/var/lib/kubelet/pods/af7c90c2-09e5-4f9d-8fa2-81dce6863688/volumes" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.233341 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" event={"ID":"e2bdda06-6da0-49de-84e5-f2d34ff27a29","Type":"ContainerStarted","Data":"814c9c1fe7facac4ba3d15c6c6ce8d1b6c4016a89f4c063194d73ad59ad25c9b"} Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.233726 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.235631 4989 generic.go:334] "Generic (PLEG): container finished" podID="1af1e529-7c4b-4ccc-94ac-2143a4ce014e" containerID="89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57" exitCode=0 Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.236368 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" event={"ID":"1af1e529-7c4b-4ccc-94ac-2143a4ce014e","Type":"ContainerDied","Data":"89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57"} Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.236393 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.236404 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt" event={"ID":"1af1e529-7c4b-4ccc-94ac-2143a4ce014e","Type":"ContainerDied","Data":"78528becf881a19400b383d080f770f91dc50d30673732e7d0ec6d14e77135f6"} Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.236435 4989 scope.go:117] "RemoveContainer" containerID="89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.242379 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.254409 4989 scope.go:117] "RemoveContainer" containerID="89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57" Dec 13 17:19:40 crc kubenswrapper[4989]: E1213 17:19:40.254880 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57\": container with ID starting with 89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57 not found: ID does not exist" containerID="89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.254933 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57"} err="failed to get container status \"89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57\": rpc error: code = NotFound desc = could not find container \"89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57\": container with ID starting with 89c14295e3cdfef5fc2f68bd77b547dbbda3a10dd8b52bf96bf41c7501ff5d57 not found: ID does not exist" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.268492 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" podStartSLOduration=4.268467889 podStartE2EDuration="4.268467889s" podCreationTimestamp="2025-12-13 17:19:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:19:40.261774749 +0000 UTC m=+194.868221907" watchObservedRunningTime="2025-12-13 17:19:40.268467889 +0000 UTC m=+194.874915027" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.348992 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt"] Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.351271 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-qvqwt"] Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.839929 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 13 17:19:40 crc kubenswrapper[4989]: E1213 17:19:40.840477 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af1e529-7c4b-4ccc-94ac-2143a4ce014e" containerName="route-controller-manager" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.840534 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af1e529-7c4b-4ccc-94ac-2143a4ce014e" containerName="route-controller-manager" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.841226 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af1e529-7c4b-4ccc-94ac-2143a4ce014e" containerName="route-controller-manager" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.842331 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.845875 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.846263 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.848861 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.918867 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 17:19:40 crc kubenswrapper[4989]: I1213 17:19:40.918916 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.016018 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5"] Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.019062 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.020123 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.020195 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.020351 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.021224 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.021331 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5"] Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.021884 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.024262 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.025000 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.025840 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.025983 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.046446 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.121116 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70816b13-44be-4e5b-aec7-a745c415d6ce-serving-cert\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.121276 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-client-ca\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.121416 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvv59\" (UniqueName: \"kubernetes.io/projected/70816b13-44be-4e5b-aec7-a745c415d6ce-kube-api-access-cvv59\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.121526 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-config\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.173530 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.222766 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvv59\" (UniqueName: \"kubernetes.io/projected/70816b13-44be-4e5b-aec7-a745c415d6ce-kube-api-access-cvv59\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.222851 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-config\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.222894 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70816b13-44be-4e5b-aec7-a745c415d6ce-serving-cert\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.222921 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-client-ca\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.223781 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-client-ca\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.225297 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-config\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.228608 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70816b13-44be-4e5b-aec7-a745c415d6ce-serving-cert\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.243105 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvv59\" (UniqueName: \"kubernetes.io/projected/70816b13-44be-4e5b-aec7-a745c415d6ce-kube-api-access-cvv59\") pod \"route-controller-manager-65d996dd87-lb8l5\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.334679 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.595995 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 13 17:19:41 crc kubenswrapper[4989]: W1213 17:19:41.603491 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poddb3c8143_2bb7_4ab6_a3f2_284ca2fdb194.slice/crio-071537a9375ca8c93415783661b13b1655ee7168ac4a902c4bdf2fb187ee98e3 WatchSource:0}: Error finding container 071537a9375ca8c93415783661b13b1655ee7168ac4a902c4bdf2fb187ee98e3: Status 404 returned error can't find the container with id 071537a9375ca8c93415783661b13b1655ee7168ac4a902c4bdf2fb187ee98e3 Dec 13 17:19:41 crc kubenswrapper[4989]: I1213 17:19:41.752744 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5"] Dec 13 17:19:41 crc kubenswrapper[4989]: W1213 17:19:41.754343 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70816b13_44be_4e5b_aec7_a745c415d6ce.slice/crio-deb7c1257fa59ac654148c756980abd3631094538cbc4e123344166df0e5dc1a WatchSource:0}: Error finding container deb7c1257fa59ac654148c756980abd3631094538cbc4e123344166df0e5dc1a: Status 404 returned error can't find the container with id deb7c1257fa59ac654148c756980abd3631094538cbc4e123344166df0e5dc1a Dec 13 17:19:42 crc kubenswrapper[4989]: I1213 17:19:42.022725 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af1e529-7c4b-4ccc-94ac-2143a4ce014e" path="/var/lib/kubelet/pods/1af1e529-7c4b-4ccc-94ac-2143a4ce014e/volumes" Dec 13 17:19:42 crc kubenswrapper[4989]: I1213 17:19:42.248815 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" event={"ID":"70816b13-44be-4e5b-aec7-a745c415d6ce","Type":"ContainerStarted","Data":"37ba6f42b810d10541f25dffd635fdb4764bd6d4ff29e7bf62cd8182134ea8ce"} Dec 13 17:19:42 crc kubenswrapper[4989]: I1213 17:19:42.248872 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" event={"ID":"70816b13-44be-4e5b-aec7-a745c415d6ce","Type":"ContainerStarted","Data":"deb7c1257fa59ac654148c756980abd3631094538cbc4e123344166df0e5dc1a"} Dec 13 17:19:42 crc kubenswrapper[4989]: I1213 17:19:42.248896 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:42 crc kubenswrapper[4989]: I1213 17:19:42.252414 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194","Type":"ContainerStarted","Data":"194c2cea8d3634fe3f883874bbb952d828c57c9cbb8c070d984d43b2b22ddc51"} Dec 13 17:19:42 crc kubenswrapper[4989]: I1213 17:19:42.252450 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194","Type":"ContainerStarted","Data":"071537a9375ca8c93415783661b13b1655ee7168ac4a902c4bdf2fb187ee98e3"} Dec 13 17:19:42 crc kubenswrapper[4989]: I1213 17:19:42.266397 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" podStartSLOduration=6.266372363 podStartE2EDuration="6.266372363s" podCreationTimestamp="2025-12-13 17:19:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:19:42.265913398 +0000 UTC m=+196.872360536" watchObservedRunningTime="2025-12-13 17:19:42.266372363 +0000 UTC m=+196.872819511" Dec 13 17:19:42 crc kubenswrapper[4989]: I1213 17:19:42.323318 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:42 crc kubenswrapper[4989]: I1213 17:19:42.341699 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.341676507 podStartE2EDuration="2.341676507s" podCreationTimestamp="2025-12-13 17:19:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:19:42.279531396 +0000 UTC m=+196.885978554" watchObservedRunningTime="2025-12-13 17:19:42.341676507 +0000 UTC m=+196.948123645" Dec 13 17:19:43 crc kubenswrapper[4989]: I1213 17:19:43.258255 4989 generic.go:334] "Generic (PLEG): container finished" podID="db3c8143-2bb7-4ab6-a3f2-284ca2fdb194" containerID="194c2cea8d3634fe3f883874bbb952d828c57c9cbb8c070d984d43b2b22ddc51" exitCode=0 Dec 13 17:19:43 crc kubenswrapper[4989]: I1213 17:19:43.258339 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194","Type":"ContainerDied","Data":"194c2cea8d3634fe3f883874bbb952d828c57c9cbb8c070d984d43b2b22ddc51"} Dec 13 17:19:44 crc kubenswrapper[4989]: I1213 17:19:44.557373 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 17:19:44 crc kubenswrapper[4989]: I1213 17:19:44.669262 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kubelet-dir\") pod \"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194\" (UID: \"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194\") " Dec 13 17:19:44 crc kubenswrapper[4989]: I1213 17:19:44.669412 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "db3c8143-2bb7-4ab6-a3f2-284ca2fdb194" (UID: "db3c8143-2bb7-4ab6-a3f2-284ca2fdb194"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:19:44 crc kubenswrapper[4989]: I1213 17:19:44.669516 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kube-api-access\") pod \"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194\" (UID: \"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194\") " Dec 13 17:19:44 crc kubenswrapper[4989]: I1213 17:19:44.670141 4989 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:44 crc kubenswrapper[4989]: I1213 17:19:44.677884 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "db3c8143-2bb7-4ab6-a3f2-284ca2fdb194" (UID: "db3c8143-2bb7-4ab6-a3f2-284ca2fdb194"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:19:44 crc kubenswrapper[4989]: I1213 17:19:44.771486 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db3c8143-2bb7-4ab6-a3f2-284ca2fdb194-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.270233 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"db3c8143-2bb7-4ab6-a3f2-284ca2fdb194","Type":"ContainerDied","Data":"071537a9375ca8c93415783661b13b1655ee7168ac4a902c4bdf2fb187ee98e3"} Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.270476 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="071537a9375ca8c93415783661b13b1655ee7168ac4a902c4bdf2fb187ee98e3" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.270381 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.837377 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 13 17:19:45 crc kubenswrapper[4989]: E1213 17:19:45.837645 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db3c8143-2bb7-4ab6-a3f2-284ca2fdb194" containerName="pruner" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.837657 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="db3c8143-2bb7-4ab6-a3f2-284ca2fdb194" containerName="pruner" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.837757 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="db3c8143-2bb7-4ab6-a3f2-284ca2fdb194" containerName="pruner" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.839649 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.842907 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.843224 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.845556 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.887316 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-var-lock\") pod \"installer-9-crc\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.887376 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kube-api-access\") pod \"installer-9-crc\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.887426 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kubelet-dir\") pod \"installer-9-crc\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.988675 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-var-lock\") pod \"installer-9-crc\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.988814 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kube-api-access\") pod \"installer-9-crc\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.988837 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-var-lock\") pod \"installer-9-crc\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.988932 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kubelet-dir\") pod \"installer-9-crc\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:19:45 crc kubenswrapper[4989]: I1213 17:19:45.988981 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kubelet-dir\") pod \"installer-9-crc\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:19:46 crc kubenswrapper[4989]: I1213 17:19:46.010024 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kube-api-access\") pod \"installer-9-crc\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:19:46 crc kubenswrapper[4989]: I1213 17:19:46.160319 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:19:46 crc kubenswrapper[4989]: I1213 17:19:46.630091 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 13 17:19:47 crc kubenswrapper[4989]: I1213 17:19:47.291179 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"aa7cc4ba-4aff-402a-a2bd-9928394c24e5","Type":"ContainerStarted","Data":"a26c97b6dce23647e4f257ddecc35f4aacdea1ad61d4d915be34b68befe7b27a"} Dec 13 17:19:47 crc kubenswrapper[4989]: I1213 17:19:47.291636 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"aa7cc4ba-4aff-402a-a2bd-9928394c24e5","Type":"ContainerStarted","Data":"de076427cc7c0505ad569a3acf5adb0327c6d19340eb5ba6606fc25d298a826a"} Dec 13 17:19:47 crc kubenswrapper[4989]: I1213 17:19:47.293479 4989 generic.go:334] "Generic (PLEG): container finished" podID="b9b8349c-c894-46a8-a043-adebd14c0dac" containerID="75eebfd7b0c286d16a5a3c290e04b2db3506bd74436ab8e8c798187c16ada6db" exitCode=0 Dec 13 17:19:47 crc kubenswrapper[4989]: I1213 17:19:47.293517 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdfb5" event={"ID":"b9b8349c-c894-46a8-a043-adebd14c0dac","Type":"ContainerDied","Data":"75eebfd7b0c286d16a5a3c290e04b2db3506bd74436ab8e8c798187c16ada6db"} Dec 13 17:19:47 crc kubenswrapper[4989]: I1213 17:19:47.312617 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.312594196 podStartE2EDuration="2.312594196s" podCreationTimestamp="2025-12-13 17:19:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:19:47.308452907 +0000 UTC m=+201.914900045" watchObservedRunningTime="2025-12-13 17:19:47.312594196 +0000 UTC m=+201.919041334" Dec 13 17:19:48 crc kubenswrapper[4989]: I1213 17:19:48.304178 4989 generic.go:334] "Generic (PLEG): container finished" podID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerID="e34f6ae15f2817dd660f08ff87ffa0f459f213a30867bbc2529c825bf1ab463e" exitCode=0 Dec 13 17:19:48 crc kubenswrapper[4989]: I1213 17:19:48.304259 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8k9w" event={"ID":"199e44f9-fad2-4498-8c5b-685a10c97ab8","Type":"ContainerDied","Data":"e34f6ae15f2817dd660f08ff87ffa0f459f213a30867bbc2529c825bf1ab463e"} Dec 13 17:19:48 crc kubenswrapper[4989]: I1213 17:19:48.307928 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zlhq7" event={"ID":"16260c45-c803-4e92-b71d-dfb9cd5b8ea2","Type":"ContainerStarted","Data":"5f1539dd6b5d87185cf413a4417022f4e2a9582cc32f9c65896c21d01ac271ae"} Dec 13 17:19:48 crc kubenswrapper[4989]: I1213 17:19:48.312276 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdfb5" event={"ID":"b9b8349c-c894-46a8-a043-adebd14c0dac","Type":"ContainerStarted","Data":"85cbaed2c7e49ecee3d0b319e18d18cc5d7c4885dd920da5423112e85afca037"} Dec 13 17:19:48 crc kubenswrapper[4989]: I1213 17:19:48.367628 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cdfb5" podStartSLOduration=3.45430835 podStartE2EDuration="48.367611318s" podCreationTimestamp="2025-12-13 17:19:00 +0000 UTC" firstStartedPulling="2025-12-13 17:19:02.771996419 +0000 UTC m=+157.378443557" lastFinishedPulling="2025-12-13 17:19:47.685299387 +0000 UTC m=+202.291746525" observedRunningTime="2025-12-13 17:19:48.36670337 +0000 UTC m=+202.973150508" watchObservedRunningTime="2025-12-13 17:19:48.367611318 +0000 UTC m=+202.974058456" Dec 13 17:19:49 crc kubenswrapper[4989]: I1213 17:19:49.318681 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8k9w" event={"ID":"199e44f9-fad2-4498-8c5b-685a10c97ab8","Type":"ContainerStarted","Data":"ab85b2383289d5502ef3478c71aa9db066983438101a3eae96a5fed3e2a920ae"} Dec 13 17:19:49 crc kubenswrapper[4989]: I1213 17:19:49.320680 4989 generic.go:334] "Generic (PLEG): container finished" podID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerID="5f1539dd6b5d87185cf413a4417022f4e2a9582cc32f9c65896c21d01ac271ae" exitCode=0 Dec 13 17:19:49 crc kubenswrapper[4989]: I1213 17:19:49.320719 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zlhq7" event={"ID":"16260c45-c803-4e92-b71d-dfb9cd5b8ea2","Type":"ContainerDied","Data":"5f1539dd6b5d87185cf413a4417022f4e2a9582cc32f9c65896c21d01ac271ae"} Dec 13 17:19:49 crc kubenswrapper[4989]: I1213 17:19:49.338962 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d8k9w" podStartSLOduration=3.227499809 podStartE2EDuration="49.338944993s" podCreationTimestamp="2025-12-13 17:19:00 +0000 UTC" firstStartedPulling="2025-12-13 17:19:02.784530933 +0000 UTC m=+157.390978071" lastFinishedPulling="2025-12-13 17:19:48.895976107 +0000 UTC m=+203.502423255" observedRunningTime="2025-12-13 17:19:49.335981661 +0000 UTC m=+203.942428819" watchObservedRunningTime="2025-12-13 17:19:49.338944993 +0000 UTC m=+203.945392131" Dec 13 17:19:50 crc kubenswrapper[4989]: I1213 17:19:50.328707 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zlhq7" event={"ID":"16260c45-c803-4e92-b71d-dfb9cd5b8ea2","Type":"ContainerStarted","Data":"e64b649cc832ab275c1d65a709fa635c790e86e702c1a466cb5608811cd28215"} Dec 13 17:19:50 crc kubenswrapper[4989]: I1213 17:19:50.704062 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:50 crc kubenswrapper[4989]: I1213 17:19:50.704335 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:19:51 crc kubenswrapper[4989]: I1213 17:19:51.041288 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zlhq7" podStartSLOduration=5.217134937 podStartE2EDuration="48.041266186s" podCreationTimestamp="2025-12-13 17:19:03 +0000 UTC" firstStartedPulling="2025-12-13 17:19:06.901535735 +0000 UTC m=+161.507982873" lastFinishedPulling="2025-12-13 17:19:49.725666984 +0000 UTC m=+204.332114122" observedRunningTime="2025-12-13 17:19:50.348471567 +0000 UTC m=+204.954918715" watchObservedRunningTime="2025-12-13 17:19:51.041266186 +0000 UTC m=+205.647713324" Dec 13 17:19:51 crc kubenswrapper[4989]: I1213 17:19:51.334344 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5xss" event={"ID":"37118126-08b8-46bd-bde4-a4faace357a2","Type":"ContainerStarted","Data":"d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9"} Dec 13 17:19:51 crc kubenswrapper[4989]: I1213 17:19:51.494555 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:51 crc kubenswrapper[4989]: I1213 17:19:51.494641 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:51 crc kubenswrapper[4989]: I1213 17:19:51.540254 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:51 crc kubenswrapper[4989]: I1213 17:19:51.849481 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d8k9w" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerName="registry-server" probeResult="failure" output=< Dec 13 17:19:51 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Dec 13 17:19:51 crc kubenswrapper[4989]: > Dec 13 17:19:52 crc kubenswrapper[4989]: I1213 17:19:52.360152 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5fxh" event={"ID":"7eeee1b2-8ab2-4635-b805-e30362bf7582","Type":"ContainerStarted","Data":"012ab4dbe64f997d6e62969a6d8431c7c5e3ce40d77a2f1e0eef4f3ba57dd4af"} Dec 13 17:19:52 crc kubenswrapper[4989]: I1213 17:19:52.364454 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m6vx" event={"ID":"157d815b-6695-4cc3-b83a-e0816757a4c4","Type":"ContainerStarted","Data":"f84143437a9011e890921e83384a8b07edb9e8e42675cb112294b8e3eb55f261"} Dec 13 17:19:52 crc kubenswrapper[4989]: I1213 17:19:52.369755 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6f77" event={"ID":"00527548-b884-4adf-8c65-4c6ca4367d77","Type":"ContainerStarted","Data":"ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb"} Dec 13 17:19:52 crc kubenswrapper[4989]: I1213 17:19:52.373360 4989 generic.go:334] "Generic (PLEG): container finished" podID="37118126-08b8-46bd-bde4-a4faace357a2" containerID="d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9" exitCode=0 Dec 13 17:19:52 crc kubenswrapper[4989]: I1213 17:19:52.373555 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5xss" event={"ID":"37118126-08b8-46bd-bde4-a4faace357a2","Type":"ContainerDied","Data":"d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9"} Dec 13 17:19:52 crc kubenswrapper[4989]: I1213 17:19:52.424865 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:53 crc kubenswrapper[4989]: I1213 17:19:53.386056 4989 generic.go:334] "Generic (PLEG): container finished" podID="7eeee1b2-8ab2-4635-b805-e30362bf7582" containerID="012ab4dbe64f997d6e62969a6d8431c7c5e3ce40d77a2f1e0eef4f3ba57dd4af" exitCode=0 Dec 13 17:19:53 crc kubenswrapper[4989]: I1213 17:19:53.386162 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5fxh" event={"ID":"7eeee1b2-8ab2-4635-b805-e30362bf7582","Type":"ContainerDied","Data":"012ab4dbe64f997d6e62969a6d8431c7c5e3ce40d77a2f1e0eef4f3ba57dd4af"} Dec 13 17:19:53 crc kubenswrapper[4989]: I1213 17:19:53.388491 4989 generic.go:334] "Generic (PLEG): container finished" podID="157d815b-6695-4cc3-b83a-e0816757a4c4" containerID="f84143437a9011e890921e83384a8b07edb9e8e42675cb112294b8e3eb55f261" exitCode=0 Dec 13 17:19:53 crc kubenswrapper[4989]: I1213 17:19:53.388587 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m6vx" event={"ID":"157d815b-6695-4cc3-b83a-e0816757a4c4","Type":"ContainerDied","Data":"f84143437a9011e890921e83384a8b07edb9e8e42675cb112294b8e3eb55f261"} Dec 13 17:19:53 crc kubenswrapper[4989]: I1213 17:19:53.391558 4989 generic.go:334] "Generic (PLEG): container finished" podID="00527548-b884-4adf-8c65-4c6ca4367d77" containerID="ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb" exitCode=0 Dec 13 17:19:53 crc kubenswrapper[4989]: I1213 17:19:53.391626 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6f77" event={"ID":"00527548-b884-4adf-8c65-4c6ca4367d77","Type":"ContainerDied","Data":"ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb"} Dec 13 17:19:53 crc kubenswrapper[4989]: I1213 17:19:53.399082 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5xss" event={"ID":"37118126-08b8-46bd-bde4-a4faace357a2","Type":"ContainerStarted","Data":"9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3"} Dec 13 17:19:53 crc kubenswrapper[4989]: I1213 17:19:53.429545 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j5xss" podStartSLOduration=3.4578819149999998 podStartE2EDuration="50.429525805s" podCreationTimestamp="2025-12-13 17:19:03 +0000 UTC" firstStartedPulling="2025-12-13 17:19:05.891221206 +0000 UTC m=+160.497668344" lastFinishedPulling="2025-12-13 17:19:52.862865096 +0000 UTC m=+207.469312234" observedRunningTime="2025-12-13 17:19:53.42363931 +0000 UTC m=+208.030086448" watchObservedRunningTime="2025-12-13 17:19:53.429525805 +0000 UTC m=+208.035972943" Dec 13 17:19:53 crc kubenswrapper[4989]: I1213 17:19:53.899951 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:53 crc kubenswrapper[4989]: I1213 17:19:53.900009 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:19:54 crc kubenswrapper[4989]: I1213 17:19:54.292657 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:54 crc kubenswrapper[4989]: I1213 17:19:54.293147 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:19:54 crc kubenswrapper[4989]: I1213 17:19:54.405921 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tbwsb" event={"ID":"5fda1871-6df5-4256-8a90-74bb43a13290","Type":"ContainerStarted","Data":"94c5c65f53f5a6981040392429390ffcd9447873d9779d4d86cd4ef43985c586"} Dec 13 17:19:54 crc kubenswrapper[4989]: I1213 17:19:54.962922 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j5xss" podUID="37118126-08b8-46bd-bde4-a4faace357a2" containerName="registry-server" probeResult="failure" output=< Dec 13 17:19:54 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Dec 13 17:19:54 crc kubenswrapper[4989]: > Dec 13 17:19:55 crc kubenswrapper[4989]: I1213 17:19:55.340875 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zlhq7" podUID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerName="registry-server" probeResult="failure" output=< Dec 13 17:19:55 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Dec 13 17:19:55 crc kubenswrapper[4989]: > Dec 13 17:19:55 crc kubenswrapper[4989]: I1213 17:19:55.418164 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6f77" event={"ID":"00527548-b884-4adf-8c65-4c6ca4367d77","Type":"ContainerStarted","Data":"93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6"} Dec 13 17:19:55 crc kubenswrapper[4989]: I1213 17:19:55.421693 4989 generic.go:334] "Generic (PLEG): container finished" podID="5fda1871-6df5-4256-8a90-74bb43a13290" containerID="94c5c65f53f5a6981040392429390ffcd9447873d9779d4d86cd4ef43985c586" exitCode=0 Dec 13 17:19:55 crc kubenswrapper[4989]: I1213 17:19:55.421737 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tbwsb" event={"ID":"5fda1871-6df5-4256-8a90-74bb43a13290","Type":"ContainerDied","Data":"94c5c65f53f5a6981040392429390ffcd9447873d9779d4d86cd4ef43985c586"} Dec 13 17:19:55 crc kubenswrapper[4989]: I1213 17:19:55.429262 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5fxh" event={"ID":"7eeee1b2-8ab2-4635-b805-e30362bf7582","Type":"ContainerStarted","Data":"8c8c838a2017f13bfa7fc054f340177ba4d645bacd8046a3eb0cc5b7c6f589d4"} Dec 13 17:19:55 crc kubenswrapper[4989]: I1213 17:19:55.432581 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m6vx" event={"ID":"157d815b-6695-4cc3-b83a-e0816757a4c4","Type":"ContainerStarted","Data":"eb6d56fe0359e1672eeea7a90c53aad413d23165018370fb667a70ca63b0d234"} Dec 13 17:19:55 crc kubenswrapper[4989]: I1213 17:19:55.437332 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x6f77" podStartSLOduration=3.856268258 podStartE2EDuration="53.437316159s" podCreationTimestamp="2025-12-13 17:19:02 +0000 UTC" firstStartedPulling="2025-12-13 17:19:04.882548259 +0000 UTC m=+159.488995397" lastFinishedPulling="2025-12-13 17:19:54.46359616 +0000 UTC m=+209.070043298" observedRunningTime="2025-12-13 17:19:55.434756199 +0000 UTC m=+210.041203337" watchObservedRunningTime="2025-12-13 17:19:55.437316159 +0000 UTC m=+210.043763297" Dec 13 17:19:55 crc kubenswrapper[4989]: I1213 17:19:55.480983 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g5fxh" podStartSLOduration=3.711074602 podStartE2EDuration="55.48096392s" podCreationTimestamp="2025-12-13 17:19:00 +0000 UTC" firstStartedPulling="2025-12-13 17:19:02.763071439 +0000 UTC m=+157.369518577" lastFinishedPulling="2025-12-13 17:19:54.532960757 +0000 UTC m=+209.139407895" observedRunningTime="2025-12-13 17:19:55.476460598 +0000 UTC m=+210.082907736" watchObservedRunningTime="2025-12-13 17:19:55.48096392 +0000 UTC m=+210.087411058" Dec 13 17:19:55 crc kubenswrapper[4989]: I1213 17:19:55.494129 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5m6vx" podStartSLOduration=5.047154676 podStartE2EDuration="53.494111492s" podCreationTimestamp="2025-12-13 17:19:02 +0000 UTC" firstStartedPulling="2025-12-13 17:19:05.895720307 +0000 UTC m=+160.502167445" lastFinishedPulling="2025-12-13 17:19:54.342677123 +0000 UTC m=+208.949124261" observedRunningTime="2025-12-13 17:19:55.492546373 +0000 UTC m=+210.098993531" watchObservedRunningTime="2025-12-13 17:19:55.494111492 +0000 UTC m=+210.100558630" Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.045037 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cdfb5"] Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.045263 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cdfb5" podUID="b9b8349c-c894-46a8-a043-adebd14c0dac" containerName="registry-server" containerID="cri-o://85cbaed2c7e49ecee3d0b319e18d18cc5d7c4885dd920da5423112e85afca037" gracePeriod=2 Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.445123 4989 generic.go:334] "Generic (PLEG): container finished" podID="b9b8349c-c894-46a8-a043-adebd14c0dac" containerID="85cbaed2c7e49ecee3d0b319e18d18cc5d7c4885dd920da5423112e85afca037" exitCode=0 Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.445235 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdfb5" event={"ID":"b9b8349c-c894-46a8-a043-adebd14c0dac","Type":"ContainerDied","Data":"85cbaed2c7e49ecee3d0b319e18d18cc5d7c4885dd920da5423112e85afca037"} Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.451386 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tbwsb" event={"ID":"5fda1871-6df5-4256-8a90-74bb43a13290","Type":"ContainerStarted","Data":"a1656b0632dec392795049dd397c6d85cad331fd22ddd95abd97d473b043de57"} Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.480327 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tbwsb" podStartSLOduration=3.400688207 podStartE2EDuration="56.480309184s" podCreationTimestamp="2025-12-13 17:19:00 +0000 UTC" firstStartedPulling="2025-12-13 17:19:02.799940196 +0000 UTC m=+157.406387334" lastFinishedPulling="2025-12-13 17:19:55.879561173 +0000 UTC m=+210.486008311" observedRunningTime="2025-12-13 17:19:56.477248728 +0000 UTC m=+211.083695886" watchObservedRunningTime="2025-12-13 17:19:56.480309184 +0000 UTC m=+211.086756322" Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.608281 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.739883 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-utilities\") pod \"b9b8349c-c894-46a8-a043-adebd14c0dac\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.739947 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-catalog-content\") pod \"b9b8349c-c894-46a8-a043-adebd14c0dac\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.740016 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p644\" (UniqueName: \"kubernetes.io/projected/b9b8349c-c894-46a8-a043-adebd14c0dac-kube-api-access-8p644\") pod \"b9b8349c-c894-46a8-a043-adebd14c0dac\" (UID: \"b9b8349c-c894-46a8-a043-adebd14c0dac\") " Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.740740 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-utilities" (OuterVolumeSpecName: "utilities") pod "b9b8349c-c894-46a8-a043-adebd14c0dac" (UID: "b9b8349c-c894-46a8-a043-adebd14c0dac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.747075 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9b8349c-c894-46a8-a043-adebd14c0dac-kube-api-access-8p644" (OuterVolumeSpecName: "kube-api-access-8p644") pod "b9b8349c-c894-46a8-a043-adebd14c0dac" (UID: "b9b8349c-c894-46a8-a043-adebd14c0dac"). InnerVolumeSpecName "kube-api-access-8p644". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.838145 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-77fbccf6-5bj5c"] Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.838486 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" podUID="e2bdda06-6da0-49de-84e5-f2d34ff27a29" containerName="controller-manager" containerID="cri-o://814c9c1fe7facac4ba3d15c6c6ce8d1b6c4016a89f4c063194d73ad59ad25c9b" gracePeriod=30 Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.845288 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.845336 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p644\" (UniqueName: \"kubernetes.io/projected/b9b8349c-c894-46a8-a043-adebd14c0dac-kube-api-access-8p644\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.876358 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5"] Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.876710 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" podUID="70816b13-44be-4e5b-aec7-a745c415d6ce" containerName="route-controller-manager" containerID="cri-o://37ba6f42b810d10541f25dffd635fdb4764bd6d4ff29e7bf62cd8182134ea8ce" gracePeriod=30 Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.900430 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9b8349c-c894-46a8-a043-adebd14c0dac" (UID: "b9b8349c-c894-46a8-a043-adebd14c0dac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:19:56 crc kubenswrapper[4989]: I1213 17:19:56.946125 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9b8349c-c894-46a8-a043-adebd14c0dac-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:57 crc kubenswrapper[4989]: I1213 17:19:57.458127 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdfb5" event={"ID":"b9b8349c-c894-46a8-a043-adebd14c0dac","Type":"ContainerDied","Data":"41176d7be3701f61707831b317c311518fad7c53f766201000ddcf6b9f4ae019"} Dec 13 17:19:57 crc kubenswrapper[4989]: I1213 17:19:57.458176 4989 scope.go:117] "RemoveContainer" containerID="85cbaed2c7e49ecee3d0b319e18d18cc5d7c4885dd920da5423112e85afca037" Dec 13 17:19:57 crc kubenswrapper[4989]: I1213 17:19:57.458278 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdfb5" Dec 13 17:19:57 crc kubenswrapper[4989]: I1213 17:19:57.475508 4989 scope.go:117] "RemoveContainer" containerID="75eebfd7b0c286d16a5a3c290e04b2db3506bd74436ab8e8c798187c16ada6db" Dec 13 17:19:57 crc kubenswrapper[4989]: I1213 17:19:57.491556 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cdfb5"] Dec 13 17:19:57 crc kubenswrapper[4989]: I1213 17:19:57.495064 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cdfb5"] Dec 13 17:19:57 crc kubenswrapper[4989]: I1213 17:19:57.508348 4989 scope.go:117] "RemoveContainer" containerID="02d37f727b2895cc05a5584fc1eab665958baee5d544adb930be48c579ed5e75" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.031119 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9b8349c-c894-46a8-a043-adebd14c0dac" path="/var/lib/kubelet/pods/b9b8349c-c894-46a8-a043-adebd14c0dac/volumes" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.473532 4989 generic.go:334] "Generic (PLEG): container finished" podID="e2bdda06-6da0-49de-84e5-f2d34ff27a29" containerID="814c9c1fe7facac4ba3d15c6c6ce8d1b6c4016a89f4c063194d73ad59ad25c9b" exitCode=0 Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.473831 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" event={"ID":"e2bdda06-6da0-49de-84e5-f2d34ff27a29","Type":"ContainerDied","Data":"814c9c1fe7facac4ba3d15c6c6ce8d1b6c4016a89f4c063194d73ad59ad25c9b"} Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.477891 4989 generic.go:334] "Generic (PLEG): container finished" podID="70816b13-44be-4e5b-aec7-a745c415d6ce" containerID="37ba6f42b810d10541f25dffd635fdb4764bd6d4ff29e7bf62cd8182134ea8ce" exitCode=0 Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.478032 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" event={"ID":"70816b13-44be-4e5b-aec7-a745c415d6ce","Type":"ContainerDied","Data":"37ba6f42b810d10541f25dffd635fdb4764bd6d4ff29e7bf62cd8182134ea8ce"} Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.676009 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.703724 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm"] Dec 13 17:19:58 crc kubenswrapper[4989]: E1213 17:19:58.703987 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70816b13-44be-4e5b-aec7-a745c415d6ce" containerName="route-controller-manager" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.704009 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="70816b13-44be-4e5b-aec7-a745c415d6ce" containerName="route-controller-manager" Dec 13 17:19:58 crc kubenswrapper[4989]: E1213 17:19:58.704033 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b8349c-c894-46a8-a043-adebd14c0dac" containerName="extract-utilities" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.704041 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b8349c-c894-46a8-a043-adebd14c0dac" containerName="extract-utilities" Dec 13 17:19:58 crc kubenswrapper[4989]: E1213 17:19:58.704057 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b8349c-c894-46a8-a043-adebd14c0dac" containerName="registry-server" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.704066 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b8349c-c894-46a8-a043-adebd14c0dac" containerName="registry-server" Dec 13 17:19:58 crc kubenswrapper[4989]: E1213 17:19:58.704079 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b8349c-c894-46a8-a043-adebd14c0dac" containerName="extract-content" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.704087 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b8349c-c894-46a8-a043-adebd14c0dac" containerName="extract-content" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.704210 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="70816b13-44be-4e5b-aec7-a745c415d6ce" containerName="route-controller-manager" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.704229 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9b8349c-c894-46a8-a043-adebd14c0dac" containerName="registry-server" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.704707 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.719137 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm"] Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.734359 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.778747 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70816b13-44be-4e5b-aec7-a745c415d6ce-serving-cert\") pod \"70816b13-44be-4e5b-aec7-a745c415d6ce\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.779243 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvv59\" (UniqueName: \"kubernetes.io/projected/70816b13-44be-4e5b-aec7-a745c415d6ce-kube-api-access-cvv59\") pod \"70816b13-44be-4e5b-aec7-a745c415d6ce\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.779283 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-client-ca\") pod \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.779309 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-proxy-ca-bundles\") pod \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.779338 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-client-ca\") pod \"70816b13-44be-4e5b-aec7-a745c415d6ce\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.779363 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vx7b\" (UniqueName: \"kubernetes.io/projected/e2bdda06-6da0-49de-84e5-f2d34ff27a29-kube-api-access-8vx7b\") pod \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.779389 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-config\") pod \"70816b13-44be-4e5b-aec7-a745c415d6ce\" (UID: \"70816b13-44be-4e5b-aec7-a745c415d6ce\") " Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.779421 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-config\") pod \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.779562 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-client-ca\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.779606 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9q56\" (UniqueName: \"kubernetes.io/projected/0bf1a5ab-fbc1-4169-893e-0117b4931068-kube-api-access-t9q56\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.779658 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0bf1a5ab-fbc1-4169-893e-0117b4931068-serving-cert\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.779680 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-config\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.780818 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-config" (OuterVolumeSpecName: "config") pod "70816b13-44be-4e5b-aec7-a745c415d6ce" (UID: "70816b13-44be-4e5b-aec7-a745c415d6ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.781498 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-client-ca" (OuterVolumeSpecName: "client-ca") pod "e2bdda06-6da0-49de-84e5-f2d34ff27a29" (UID: "e2bdda06-6da0-49de-84e5-f2d34ff27a29"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.782021 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e2bdda06-6da0-49de-84e5-f2d34ff27a29" (UID: "e2bdda06-6da0-49de-84e5-f2d34ff27a29"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.782471 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-client-ca" (OuterVolumeSpecName: "client-ca") pod "70816b13-44be-4e5b-aec7-a745c415d6ce" (UID: "70816b13-44be-4e5b-aec7-a745c415d6ce"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.782682 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-config" (OuterVolumeSpecName: "config") pod "e2bdda06-6da0-49de-84e5-f2d34ff27a29" (UID: "e2bdda06-6da0-49de-84e5-f2d34ff27a29"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.784457 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70816b13-44be-4e5b-aec7-a745c415d6ce-kube-api-access-cvv59" (OuterVolumeSpecName: "kube-api-access-cvv59") pod "70816b13-44be-4e5b-aec7-a745c415d6ce" (UID: "70816b13-44be-4e5b-aec7-a745c415d6ce"). InnerVolumeSpecName "kube-api-access-cvv59". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.784579 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70816b13-44be-4e5b-aec7-a745c415d6ce-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "70816b13-44be-4e5b-aec7-a745c415d6ce" (UID: "70816b13-44be-4e5b-aec7-a745c415d6ce"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.784677 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2bdda06-6da0-49de-84e5-f2d34ff27a29-kube-api-access-8vx7b" (OuterVolumeSpecName: "kube-api-access-8vx7b") pod "e2bdda06-6da0-49de-84e5-f2d34ff27a29" (UID: "e2bdda06-6da0-49de-84e5-f2d34ff27a29"). InnerVolumeSpecName "kube-api-access-8vx7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880246 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2bdda06-6da0-49de-84e5-f2d34ff27a29-serving-cert\") pod \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\" (UID: \"e2bdda06-6da0-49de-84e5-f2d34ff27a29\") " Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880576 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-client-ca\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880623 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9q56\" (UniqueName: \"kubernetes.io/projected/0bf1a5ab-fbc1-4169-893e-0117b4931068-kube-api-access-t9q56\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880665 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0bf1a5ab-fbc1-4169-893e-0117b4931068-serving-cert\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880707 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-config\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880779 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880806 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vx7b\" (UniqueName: \"kubernetes.io/projected/e2bdda06-6da0-49de-84e5-f2d34ff27a29-kube-api-access-8vx7b\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880819 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70816b13-44be-4e5b-aec7-a745c415d6ce-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880828 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880837 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70816b13-44be-4e5b-aec7-a745c415d6ce-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880848 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvv59\" (UniqueName: \"kubernetes.io/projected/70816b13-44be-4e5b-aec7-a745c415d6ce-kube-api-access-cvv59\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880857 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.880865 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2bdda06-6da0-49de-84e5-f2d34ff27a29-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.881627 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-client-ca\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.881937 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-config\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.884922 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2bdda06-6da0-49de-84e5-f2d34ff27a29-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e2bdda06-6da0-49de-84e5-f2d34ff27a29" (UID: "e2bdda06-6da0-49de-84e5-f2d34ff27a29"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.885192 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0bf1a5ab-fbc1-4169-893e-0117b4931068-serving-cert\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.897100 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9q56\" (UniqueName: \"kubernetes.io/projected/0bf1a5ab-fbc1-4169-893e-0117b4931068-kube-api-access-t9q56\") pod \"route-controller-manager-85f797b5f-pwjpm\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:58 crc kubenswrapper[4989]: I1213 17:19:58.981743 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2bdda06-6da0-49de-84e5-f2d34ff27a29-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.065074 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.293077 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm"] Dec 13 17:19:59 crc kubenswrapper[4989]: W1213 17:19:59.297310 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0bf1a5ab_fbc1_4169_893e_0117b4931068.slice/crio-db617a880ea432d16e200b0528c12276309ed2ddebb25df01a4222c2beaf4fdb WatchSource:0}: Error finding container db617a880ea432d16e200b0528c12276309ed2ddebb25df01a4222c2beaf4fdb: Status 404 returned error can't find the container with id db617a880ea432d16e200b0528c12276309ed2ddebb25df01a4222c2beaf4fdb Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.489509 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" event={"ID":"0bf1a5ab-fbc1-4169-893e-0117b4931068","Type":"ContainerStarted","Data":"db617a880ea432d16e200b0528c12276309ed2ddebb25df01a4222c2beaf4fdb"} Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.491269 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.491256 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" event={"ID":"e2bdda06-6da0-49de-84e5-f2d34ff27a29","Type":"ContainerDied","Data":"2c45facf861b25f5c184ff376131589b69a1d01f52b022ff6d9653f2a7d3bafc"} Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.491436 4989 scope.go:117] "RemoveContainer" containerID="814c9c1fe7facac4ba3d15c6c6ce8d1b6c4016a89f4c063194d73ad59ad25c9b" Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.493276 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.493241 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5" event={"ID":"70816b13-44be-4e5b-aec7-a745c415d6ce","Type":"ContainerDied","Data":"deb7c1257fa59ac654148c756980abd3631094538cbc4e123344166df0e5dc1a"} Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.496404 4989 patch_prober.go:28] interesting pod/controller-manager-77fbccf6-5bj5c container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.496469 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-77fbccf6-5bj5c" podUID="e2bdda06-6da0-49de-84e5-f2d34ff27a29" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.506512 4989 scope.go:117] "RemoveContainer" containerID="37ba6f42b810d10541f25dffd635fdb4764bd6d4ff29e7bf62cd8182134ea8ce" Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.522289 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-77fbccf6-5bj5c"] Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.526750 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-77fbccf6-5bj5c"] Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.540578 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5"] Dec 13 17:19:59 crc kubenswrapper[4989]: I1213 17:19:59.545737 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65d996dd87-lb8l5"] Dec 13 17:20:00 crc kubenswrapper[4989]: I1213 17:20:00.022673 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70816b13-44be-4e5b-aec7-a745c415d6ce" path="/var/lib/kubelet/pods/70816b13-44be-4e5b-aec7-a745c415d6ce/volumes" Dec 13 17:20:00 crc kubenswrapper[4989]: I1213 17:20:00.023409 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2bdda06-6da0-49de-84e5-f2d34ff27a29" path="/var/lib/kubelet/pods/e2bdda06-6da0-49de-84e5-f2d34ff27a29/volumes" Dec 13 17:20:00 crc kubenswrapper[4989]: I1213 17:20:00.502200 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" event={"ID":"0bf1a5ab-fbc1-4169-893e-0117b4931068","Type":"ContainerStarted","Data":"daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22"} Dec 13 17:20:00 crc kubenswrapper[4989]: I1213 17:20:00.503171 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:20:00 crc kubenswrapper[4989]: I1213 17:20:00.507818 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:20:00 crc kubenswrapper[4989]: I1213 17:20:00.520085 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" podStartSLOduration=4.52006539 podStartE2EDuration="4.52006539s" podCreationTimestamp="2025-12-13 17:19:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:20:00.519619376 +0000 UTC m=+215.126066514" watchObservedRunningTime="2025-12-13 17:20:00.52006539 +0000 UTC m=+215.126512528" Dec 13 17:20:00 crc kubenswrapper[4989]: I1213 17:20:00.745635 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:20:00 crc kubenswrapper[4989]: I1213 17:20:00.785497 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.027496 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr"] Dec 13 17:20:01 crc kubenswrapper[4989]: E1213 17:20:01.027708 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2bdda06-6da0-49de-84e5-f2d34ff27a29" containerName="controller-manager" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.027721 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2bdda06-6da0-49de-84e5-f2d34ff27a29" containerName="controller-manager" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.027841 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2bdda06-6da0-49de-84e5-f2d34ff27a29" containerName="controller-manager" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.028269 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.030454 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.031305 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.031443 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.031574 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.031671 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.031764 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.034954 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr"] Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.040402 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.195729 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.195842 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.216090 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-proxy-ca-bundles\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.216144 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-config\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.216192 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-client-ca\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.216243 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnvlx\" (UniqueName: \"kubernetes.io/projected/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-kube-api-access-fnvlx\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.216273 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-serving-cert\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.232631 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.317019 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-proxy-ca-bundles\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.317072 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-config\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.317141 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-client-ca\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.317168 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnvlx\" (UniqueName: \"kubernetes.io/projected/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-kube-api-access-fnvlx\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.317187 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-serving-cert\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.318150 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-client-ca\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.318751 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-config\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.318763 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-proxy-ca-bundles\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.328162 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-serving-cert\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.334967 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnvlx\" (UniqueName: \"kubernetes.io/projected/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-kube-api-access-fnvlx\") pod \"controller-manager-6f649d8b6b-dmbxr\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.348763 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.423967 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.424016 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.475669 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.551160 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.554820 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:20:01 crc kubenswrapper[4989]: I1213 17:20:01.556927 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr"] Dec 13 17:20:02 crc kubenswrapper[4989]: I1213 17:20:02.513724 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" event={"ID":"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb","Type":"ContainerStarted","Data":"2803d22a6271fe72720df3874243c97da3c31c1fa5629c3188d4a02f990bb935"} Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.210773 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.211400 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.266554 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.339607 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.339678 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.389354 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.445945 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g5fxh"] Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.518694 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g5fxh" podUID="7eeee1b2-8ab2-4635-b805-e30362bf7582" containerName="registry-server" containerID="cri-o://8c8c838a2017f13bfa7fc054f340177ba4d645bacd8046a3eb0cc5b7c6f589d4" gracePeriod=2 Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.570743 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.574349 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.941971 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:20:03 crc kubenswrapper[4989]: I1213 17:20:03.982451 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:20:04 crc kubenswrapper[4989]: I1213 17:20:04.352768 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:20:04 crc kubenswrapper[4989]: I1213 17:20:04.420345 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:20:05 crc kubenswrapper[4989]: I1213 17:20:05.546286 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" event={"ID":"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb","Type":"ContainerStarted","Data":"5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348"} Dec 13 17:20:05 crc kubenswrapper[4989]: I1213 17:20:05.546610 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:05 crc kubenswrapper[4989]: I1213 17:20:05.547966 4989 patch_prober.go:28] interesting pod/controller-manager-6f649d8b6b-dmbxr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" start-of-body= Dec 13 17:20:05 crc kubenswrapper[4989]: I1213 17:20:05.548026 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" podUID="b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" Dec 13 17:20:05 crc kubenswrapper[4989]: I1213 17:20:05.567238 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" podStartSLOduration=9.567211683 podStartE2EDuration="9.567211683s" podCreationTimestamp="2025-12-13 17:19:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:20:05.563551668 +0000 UTC m=+220.169998836" watchObservedRunningTime="2025-12-13 17:20:05.567211683 +0000 UTC m=+220.173658821" Dec 13 17:20:05 crc kubenswrapper[4989]: I1213 17:20:05.647471 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m6vx"] Dec 13 17:20:05 crc kubenswrapper[4989]: I1213 17:20:05.647943 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5m6vx" podUID="157d815b-6695-4cc3-b83a-e0816757a4c4" containerName="registry-server" containerID="cri-o://eb6d56fe0359e1672eeea7a90c53aad413d23165018370fb667a70ca63b0d234" gracePeriod=2 Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.553620 4989 generic.go:334] "Generic (PLEG): container finished" podID="157d815b-6695-4cc3-b83a-e0816757a4c4" containerID="eb6d56fe0359e1672eeea7a90c53aad413d23165018370fb667a70ca63b0d234" exitCode=0 Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.553703 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m6vx" event={"ID":"157d815b-6695-4cc3-b83a-e0816757a4c4","Type":"ContainerDied","Data":"eb6d56fe0359e1672eeea7a90c53aad413d23165018370fb667a70ca63b0d234"} Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.554979 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5fxh_7eeee1b2-8ab2-4635-b805-e30362bf7582/registry-server/0.log" Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.555576 4989 generic.go:334] "Generic (PLEG): container finished" podID="7eeee1b2-8ab2-4635-b805-e30362bf7582" containerID="8c8c838a2017f13bfa7fc054f340177ba4d645bacd8046a3eb0cc5b7c6f589d4" exitCode=137 Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.555615 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5fxh" event={"ID":"7eeee1b2-8ab2-4635-b805-e30362bf7582","Type":"ContainerDied","Data":"8c8c838a2017f13bfa7fc054f340177ba4d645bacd8046a3eb0cc5b7c6f589d4"} Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.559300 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.846633 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5fxh_7eeee1b2-8ab2-4635-b805-e30362bf7582/registry-server/0.log" Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.847900 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.909294 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-utilities\") pod \"7eeee1b2-8ab2-4635-b805-e30362bf7582\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.909418 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpz9q\" (UniqueName: \"kubernetes.io/projected/7eeee1b2-8ab2-4635-b805-e30362bf7582-kube-api-access-fpz9q\") pod \"7eeee1b2-8ab2-4635-b805-e30362bf7582\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.909566 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-catalog-content\") pod \"7eeee1b2-8ab2-4635-b805-e30362bf7582\" (UID: \"7eeee1b2-8ab2-4635-b805-e30362bf7582\") " Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.910482 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-utilities" (OuterVolumeSpecName: "utilities") pod "7eeee1b2-8ab2-4635-b805-e30362bf7582" (UID: "7eeee1b2-8ab2-4635-b805-e30362bf7582"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.915811 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eeee1b2-8ab2-4635-b805-e30362bf7582-kube-api-access-fpz9q" (OuterVolumeSpecName: "kube-api-access-fpz9q") pod "7eeee1b2-8ab2-4635-b805-e30362bf7582" (UID: "7eeee1b2-8ab2-4635-b805-e30362bf7582"). InnerVolumeSpecName "kube-api-access-fpz9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:20:06 crc kubenswrapper[4989]: I1213 17:20:06.960953 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7eeee1b2-8ab2-4635-b805-e30362bf7582" (UID: "7eeee1b2-8ab2-4635-b805-e30362bf7582"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.011859 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.011901 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpz9q\" (UniqueName: \"kubernetes.io/projected/7eeee1b2-8ab2-4635-b805-e30362bf7582-kube-api-access-fpz9q\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.011914 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7eeee1b2-8ab2-4635-b805-e30362bf7582-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.250643 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.316473 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-utilities\") pod \"157d815b-6695-4cc3-b83a-e0816757a4c4\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.316576 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xjk8\" (UniqueName: \"kubernetes.io/projected/157d815b-6695-4cc3-b83a-e0816757a4c4-kube-api-access-6xjk8\") pod \"157d815b-6695-4cc3-b83a-e0816757a4c4\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.316680 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-catalog-content\") pod \"157d815b-6695-4cc3-b83a-e0816757a4c4\" (UID: \"157d815b-6695-4cc3-b83a-e0816757a4c4\") " Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.317241 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-utilities" (OuterVolumeSpecName: "utilities") pod "157d815b-6695-4cc3-b83a-e0816757a4c4" (UID: "157d815b-6695-4cc3-b83a-e0816757a4c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.320026 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/157d815b-6695-4cc3-b83a-e0816757a4c4-kube-api-access-6xjk8" (OuterVolumeSpecName: "kube-api-access-6xjk8") pod "157d815b-6695-4cc3-b83a-e0816757a4c4" (UID: "157d815b-6695-4cc3-b83a-e0816757a4c4"). InnerVolumeSpecName "kube-api-access-6xjk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.336294 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "157d815b-6695-4cc3-b83a-e0816757a4c4" (UID: "157d815b-6695-4cc3-b83a-e0816757a4c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.418431 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.418475 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/157d815b-6695-4cc3-b83a-e0816757a4c4-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.418490 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xjk8\" (UniqueName: \"kubernetes.io/projected/157d815b-6695-4cc3-b83a-e0816757a4c4-kube-api-access-6xjk8\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.562374 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5fxh_7eeee1b2-8ab2-4635-b805-e30362bf7582/registry-server/0.log" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.563302 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5fxh" event={"ID":"7eeee1b2-8ab2-4635-b805-e30362bf7582","Type":"ContainerDied","Data":"89e3bbab87fd6d8e3750f48cddd058d1746db4f5ce25b9f9e99a4244600bd3de"} Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.563350 4989 scope.go:117] "RemoveContainer" containerID="8c8c838a2017f13bfa7fc054f340177ba4d645bacd8046a3eb0cc5b7c6f589d4" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.563478 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5fxh" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.568001 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m6vx" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.567994 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m6vx" event={"ID":"157d815b-6695-4cc3-b83a-e0816757a4c4","Type":"ContainerDied","Data":"bd460960bab9a2d211bbaac5c9040acf33d384747720dc93eab05f8f6b6b1720"} Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.585837 4989 scope.go:117] "RemoveContainer" containerID="012ab4dbe64f997d6e62969a6d8431c7c5e3ce40d77a2f1e0eef4f3ba57dd4af" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.599657 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g5fxh"] Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.608813 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g5fxh"] Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.612621 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m6vx"] Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.615099 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m6vx"] Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.623854 4989 scope.go:117] "RemoveContainer" containerID="15f119ecc45725dc4e04dd710af40f456be116658724b37554071547e1c8ac2c" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.637391 4989 scope.go:117] "RemoveContainer" containerID="eb6d56fe0359e1672eeea7a90c53aad413d23165018370fb667a70ca63b0d234" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.649474 4989 scope.go:117] "RemoveContainer" containerID="f84143437a9011e890921e83384a8b07edb9e8e42675cb112294b8e3eb55f261" Dec 13 17:20:07 crc kubenswrapper[4989]: I1213 17:20:07.661302 4989 scope.go:117] "RemoveContainer" containerID="a7441b88eb35a3f756f16f2d9eaa631a92a5e28c10f391f6f7d487b6a4c56379" Dec 13 17:20:08 crc kubenswrapper[4989]: I1213 17:20:08.022576 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="157d815b-6695-4cc3-b83a-e0816757a4c4" path="/var/lib/kubelet/pods/157d815b-6695-4cc3-b83a-e0816757a4c4/volumes" Dec 13 17:20:08 crc kubenswrapper[4989]: I1213 17:20:08.024454 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eeee1b2-8ab2-4635-b805-e30362bf7582" path="/var/lib/kubelet/pods/7eeee1b2-8ab2-4635-b805-e30362bf7582/volumes" Dec 13 17:20:08 crc kubenswrapper[4989]: I1213 17:20:08.055337 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zlhq7"] Dec 13 17:20:08 crc kubenswrapper[4989]: I1213 17:20:08.055835 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zlhq7" podUID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerName="registry-server" containerID="cri-o://e64b649cc832ab275c1d65a709fa635c790e86e702c1a466cb5608811cd28215" gracePeriod=2 Dec 13 17:20:08 crc kubenswrapper[4989]: I1213 17:20:08.829744 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:20:08 crc kubenswrapper[4989]: I1213 17:20:08.829860 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:20:08 crc kubenswrapper[4989]: I1213 17:20:08.829930 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:20:08 crc kubenswrapper[4989]: I1213 17:20:08.831010 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0"} pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 17:20:08 crc kubenswrapper[4989]: I1213 17:20:08.831118 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" containerID="cri-o://8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0" gracePeriod=600 Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.584193 4989 generic.go:334] "Generic (PLEG): container finished" podID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerID="8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0" exitCode=0 Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.584507 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerDied","Data":"8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0"} Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.591409 4989 generic.go:334] "Generic (PLEG): container finished" podID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerID="e64b649cc832ab275c1d65a709fa635c790e86e702c1a466cb5608811cd28215" exitCode=0 Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.591453 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zlhq7" event={"ID":"16260c45-c803-4e92-b71d-dfb9cd5b8ea2","Type":"ContainerDied","Data":"e64b649cc832ab275c1d65a709fa635c790e86e702c1a466cb5608811cd28215"} Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.591515 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zlhq7" event={"ID":"16260c45-c803-4e92-b71d-dfb9cd5b8ea2","Type":"ContainerDied","Data":"bd1d861b2adfe9d57e327665f36f9a2d68cf4e1fe16a62c1c7e64d60994c5f13"} Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.591530 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd1d861b2adfe9d57e327665f36f9a2d68cf4e1fe16a62c1c7e64d60994c5f13" Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.593114 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.652860 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-catalog-content\") pod \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.653314 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdd6v\" (UniqueName: \"kubernetes.io/projected/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-kube-api-access-fdd6v\") pod \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.653391 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-utilities\") pod \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\" (UID: \"16260c45-c803-4e92-b71d-dfb9cd5b8ea2\") " Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.654366 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-utilities" (OuterVolumeSpecName: "utilities") pod "16260c45-c803-4e92-b71d-dfb9cd5b8ea2" (UID: "16260c45-c803-4e92-b71d-dfb9cd5b8ea2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.675553 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-kube-api-access-fdd6v" (OuterVolumeSpecName: "kube-api-access-fdd6v") pod "16260c45-c803-4e92-b71d-dfb9cd5b8ea2" (UID: "16260c45-c803-4e92-b71d-dfb9cd5b8ea2"). InnerVolumeSpecName "kube-api-access-fdd6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.754426 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdd6v\" (UniqueName: \"kubernetes.io/projected/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-kube-api-access-fdd6v\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.754456 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.778485 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16260c45-c803-4e92-b71d-dfb9cd5b8ea2" (UID: "16260c45-c803-4e92-b71d-dfb9cd5b8ea2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:20:09 crc kubenswrapper[4989]: I1213 17:20:09.856139 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16260c45-c803-4e92-b71d-dfb9cd5b8ea2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:10 crc kubenswrapper[4989]: I1213 17:20:10.600809 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zlhq7" Dec 13 17:20:10 crc kubenswrapper[4989]: I1213 17:20:10.600778 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerStarted","Data":"1707457e30b3e0348e549e5cd99a4006766e7e37c627bd9d4764bc772d5280c9"} Dec 13 17:20:10 crc kubenswrapper[4989]: I1213 17:20:10.634547 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zlhq7"] Dec 13 17:20:10 crc kubenswrapper[4989]: I1213 17:20:10.637918 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zlhq7"] Dec 13 17:20:11 crc kubenswrapper[4989]: I1213 17:20:11.611437 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dvtmv"] Dec 13 17:20:12 crc kubenswrapper[4989]: I1213 17:20:12.025228 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" path="/var/lib/kubelet/pods/16260c45-c803-4e92-b71d-dfb9cd5b8ea2/volumes" Dec 13 17:20:16 crc kubenswrapper[4989]: I1213 17:20:16.809854 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr"] Dec 13 17:20:16 crc kubenswrapper[4989]: I1213 17:20:16.811162 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" podUID="b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" containerName="controller-manager" containerID="cri-o://5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348" gracePeriod=30 Dec 13 17:20:16 crc kubenswrapper[4989]: I1213 17:20:16.912029 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm"] Dec 13 17:20:16 crc kubenswrapper[4989]: I1213 17:20:16.912276 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" podUID="0bf1a5ab-fbc1-4169-893e-0117b4931068" containerName="route-controller-manager" containerID="cri-o://daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22" gracePeriod=30 Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.451218 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.500006 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.570580 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-config\") pod \"0bf1a5ab-fbc1-4169-893e-0117b4931068\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.570631 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-serving-cert\") pod \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.570697 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9q56\" (UniqueName: \"kubernetes.io/projected/0bf1a5ab-fbc1-4169-893e-0117b4931068-kube-api-access-t9q56\") pod \"0bf1a5ab-fbc1-4169-893e-0117b4931068\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.570714 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-proxy-ca-bundles\") pod \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.570748 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnvlx\" (UniqueName: \"kubernetes.io/projected/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-kube-api-access-fnvlx\") pod \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.570778 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-client-ca\") pod \"0bf1a5ab-fbc1-4169-893e-0117b4931068\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.570805 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0bf1a5ab-fbc1-4169-893e-0117b4931068-serving-cert\") pod \"0bf1a5ab-fbc1-4169-893e-0117b4931068\" (UID: \"0bf1a5ab-fbc1-4169-893e-0117b4931068\") " Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.570832 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-config\") pod \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.570848 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-client-ca\") pod \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\" (UID: \"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb\") " Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.571608 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-config" (OuterVolumeSpecName: "config") pod "0bf1a5ab-fbc1-4169-893e-0117b4931068" (UID: "0bf1a5ab-fbc1-4169-893e-0117b4931068"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.571752 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-client-ca" (OuterVolumeSpecName: "client-ca") pod "b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" (UID: "b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.572140 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-client-ca" (OuterVolumeSpecName: "client-ca") pod "0bf1a5ab-fbc1-4169-893e-0117b4931068" (UID: "0bf1a5ab-fbc1-4169-893e-0117b4931068"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.573004 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-config" (OuterVolumeSpecName: "config") pod "b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" (UID: "b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.573335 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" (UID: "b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.577411 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf1a5ab-fbc1-4169-893e-0117b4931068-kube-api-access-t9q56" (OuterVolumeSpecName: "kube-api-access-t9q56") pod "0bf1a5ab-fbc1-4169-893e-0117b4931068" (UID: "0bf1a5ab-fbc1-4169-893e-0117b4931068"). InnerVolumeSpecName "kube-api-access-t9q56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.577894 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf1a5ab-fbc1-4169-893e-0117b4931068-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0bf1a5ab-fbc1-4169-893e-0117b4931068" (UID: "0bf1a5ab-fbc1-4169-893e-0117b4931068"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.578349 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-kube-api-access-fnvlx" (OuterVolumeSpecName: "kube-api-access-fnvlx") pod "b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" (UID: "b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb"). InnerVolumeSpecName "kube-api-access-fnvlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.578919 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" (UID: "b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.638086 4989 generic.go:334] "Generic (PLEG): container finished" podID="b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" containerID="5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348" exitCode=0 Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.638180 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" event={"ID":"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb","Type":"ContainerDied","Data":"5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348"} Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.638191 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.638212 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr" event={"ID":"b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb","Type":"ContainerDied","Data":"2803d22a6271fe72720df3874243c97da3c31c1fa5629c3188d4a02f990bb935"} Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.638236 4989 scope.go:117] "RemoveContainer" containerID="5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.648742 4989 generic.go:334] "Generic (PLEG): container finished" podID="0bf1a5ab-fbc1-4169-893e-0117b4931068" containerID="daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22" exitCode=0 Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.648820 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" event={"ID":"0bf1a5ab-fbc1-4169-893e-0117b4931068","Type":"ContainerDied","Data":"daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22"} Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.648853 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" event={"ID":"0bf1a5ab-fbc1-4169-893e-0117b4931068","Type":"ContainerDied","Data":"db617a880ea432d16e200b0528c12276309ed2ddebb25df01a4222c2beaf4fdb"} Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.648916 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.666878 4989 scope.go:117] "RemoveContainer" containerID="5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348" Dec 13 17:20:17 crc kubenswrapper[4989]: E1213 17:20:17.669010 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348\": container with ID starting with 5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348 not found: ID does not exist" containerID="5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.669062 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348"} err="failed to get container status \"5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348\": rpc error: code = NotFound desc = could not find container \"5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348\": container with ID starting with 5028682d4008d5f9cc8573f9f1b2f8ccc3883b3c787654197ee81773ec120348 not found: ID does not exist" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.669089 4989 scope.go:117] "RemoveContainer" containerID="daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.671883 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.671920 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.671936 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.671948 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.671960 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.671973 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9q56\" (UniqueName: \"kubernetes.io/projected/0bf1a5ab-fbc1-4169-893e-0117b4931068-kube-api-access-t9q56\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.671985 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnvlx\" (UniqueName: \"kubernetes.io/projected/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb-kube-api-access-fnvlx\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.671997 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0bf1a5ab-fbc1-4169-893e-0117b4931068-client-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.672007 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0bf1a5ab-fbc1-4169-893e-0117b4931068-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.691248 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr"] Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.693265 4989 scope.go:117] "RemoveContainer" containerID="daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22" Dec 13 17:20:17 crc kubenswrapper[4989]: E1213 17:20:17.695118 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22\": container with ID starting with daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22 not found: ID does not exist" containerID="daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.695183 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22"} err="failed to get container status \"daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22\": rpc error: code = NotFound desc = could not find container \"daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22\": container with ID starting with daebec1f9f72ddf07e53bdf4dd2ffc0d4a17b840b04ce388611d31fa6674ac22 not found: ID does not exist" Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.697258 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6f649d8b6b-dmbxr"] Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.706212 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm"] Dec 13 17:20:17 crc kubenswrapper[4989]: I1213 17:20:17.717819 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85f797b5f-pwjpm"] Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.020173 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf1a5ab-fbc1-4169-893e-0117b4931068" path="/var/lib/kubelet/pods/0bf1a5ab-fbc1-4169-893e-0117b4931068/volumes" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.020672 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" path="/var/lib/kubelet/pods/b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb/volumes" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038339 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5b6b7f6459-79dhf"] Dec 13 17:20:18 crc kubenswrapper[4989]: E1213 17:20:18.038637 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="157d815b-6695-4cc3-b83a-e0816757a4c4" containerName="extract-utilities" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038661 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="157d815b-6695-4cc3-b83a-e0816757a4c4" containerName="extract-utilities" Dec 13 17:20:18 crc kubenswrapper[4989]: E1213 17:20:18.038680 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eeee1b2-8ab2-4635-b805-e30362bf7582" containerName="extract-utilities" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038689 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eeee1b2-8ab2-4635-b805-e30362bf7582" containerName="extract-utilities" Dec 13 17:20:18 crc kubenswrapper[4989]: E1213 17:20:18.038699 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="157d815b-6695-4cc3-b83a-e0816757a4c4" containerName="extract-content" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038706 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="157d815b-6695-4cc3-b83a-e0816757a4c4" containerName="extract-content" Dec 13 17:20:18 crc kubenswrapper[4989]: E1213 17:20:18.038717 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="157d815b-6695-4cc3-b83a-e0816757a4c4" containerName="registry-server" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038724 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="157d815b-6695-4cc3-b83a-e0816757a4c4" containerName="registry-server" Dec 13 17:20:18 crc kubenswrapper[4989]: E1213 17:20:18.038732 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eeee1b2-8ab2-4635-b805-e30362bf7582" containerName="extract-content" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038741 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eeee1b2-8ab2-4635-b805-e30362bf7582" containerName="extract-content" Dec 13 17:20:18 crc kubenswrapper[4989]: E1213 17:20:18.038751 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eeee1b2-8ab2-4635-b805-e30362bf7582" containerName="registry-server" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038757 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eeee1b2-8ab2-4635-b805-e30362bf7582" containerName="registry-server" Dec 13 17:20:18 crc kubenswrapper[4989]: E1213 17:20:18.038765 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" containerName="controller-manager" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038772 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" containerName="controller-manager" Dec 13 17:20:18 crc kubenswrapper[4989]: E1213 17:20:18.038781 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerName="extract-content" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038822 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerName="extract-content" Dec 13 17:20:18 crc kubenswrapper[4989]: E1213 17:20:18.038833 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerName="extract-utilities" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038839 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerName="extract-utilities" Dec 13 17:20:18 crc kubenswrapper[4989]: E1213 17:20:18.038845 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerName="registry-server" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038851 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerName="registry-server" Dec 13 17:20:18 crc kubenswrapper[4989]: E1213 17:20:18.038859 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf1a5ab-fbc1-4169-893e-0117b4931068" containerName="route-controller-manager" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038865 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf1a5ab-fbc1-4169-893e-0117b4931068" containerName="route-controller-manager" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038962 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eeee1b2-8ab2-4635-b805-e30362bf7582" containerName="registry-server" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038974 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="16260c45-c803-4e92-b71d-dfb9cd5b8ea2" containerName="registry-server" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038983 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4190bca-3afd-4861-9c6e-4c6c0b4cd1eb" containerName="controller-manager" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038991 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf1a5ab-fbc1-4169-893e-0117b4931068" containerName="route-controller-manager" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.038997 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="157d815b-6695-4cc3-b83a-e0816757a4c4" containerName="registry-server" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.039426 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.041011 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h"] Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.041660 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.041659 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.041806 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.041860 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.041908 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.044046 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.044567 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.044768 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.045064 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.045104 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.045070 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.045385 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.048723 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.051667 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.056061 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h"] Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.059367 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b6b7f6459-79dhf"] Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.077764 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-proxy-ca-bundles\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.077847 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-serving-cert\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.077965 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-config\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.078029 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l47lj\" (UniqueName: \"kubernetes.io/projected/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-kube-api-access-l47lj\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.078058 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-client-ca\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.178742 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3d71144-6e47-4a55-baa1-d523fff194e0-config\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.178800 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-config\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.178834 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l47lj\" (UniqueName: \"kubernetes.io/projected/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-kube-api-access-l47lj\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.178852 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-client-ca\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.178870 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f3d71144-6e47-4a55-baa1-d523fff194e0-client-ca\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.178886 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sggzh\" (UniqueName: \"kubernetes.io/projected/f3d71144-6e47-4a55-baa1-d523fff194e0-kube-api-access-sggzh\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.178920 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-proxy-ca-bundles\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.178939 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3d71144-6e47-4a55-baa1-d523fff194e0-serving-cert\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.178967 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-serving-cert\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.179994 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-client-ca\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.180441 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-proxy-ca-bundles\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.180596 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-config\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.190063 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-serving-cert\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.195506 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l47lj\" (UniqueName: \"kubernetes.io/projected/be498ee8-a5fa-400a-b1e4-99e63b9eb6b4-kube-api-access-l47lj\") pod \"controller-manager-5b6b7f6459-79dhf\" (UID: \"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4\") " pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.279920 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3d71144-6e47-4a55-baa1-d523fff194e0-config\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.280009 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f3d71144-6e47-4a55-baa1-d523fff194e0-client-ca\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.280030 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sggzh\" (UniqueName: \"kubernetes.io/projected/f3d71144-6e47-4a55-baa1-d523fff194e0-kube-api-access-sggzh\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.280068 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3d71144-6e47-4a55-baa1-d523fff194e0-serving-cert\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.281153 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3d71144-6e47-4a55-baa1-d523fff194e0-config\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.281330 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f3d71144-6e47-4a55-baa1-d523fff194e0-client-ca\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.283963 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3d71144-6e47-4a55-baa1-d523fff194e0-serving-cert\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.296520 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sggzh\" (UniqueName: \"kubernetes.io/projected/f3d71144-6e47-4a55-baa1-d523fff194e0-kube-api-access-sggzh\") pod \"route-controller-manager-78ff4f79c4-lp55h\" (UID: \"f3d71144-6e47-4a55-baa1-d523fff194e0\") " pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.356669 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.364443 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.789421 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h"] Dec 13 17:20:18 crc kubenswrapper[4989]: W1213 17:20:18.794109 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3d71144_6e47_4a55_baa1_d523fff194e0.slice/crio-4cd9879590bda2d9b3ccd5052cdf1080c20eaeb226266a0d1bb2f6f0949592a0 WatchSource:0}: Error finding container 4cd9879590bda2d9b3ccd5052cdf1080c20eaeb226266a0d1bb2f6f0949592a0: Status 404 returned error can't find the container with id 4cd9879590bda2d9b3ccd5052cdf1080c20eaeb226266a0d1bb2f6f0949592a0 Dec 13 17:20:18 crc kubenswrapper[4989]: I1213 17:20:18.844758 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b6b7f6459-79dhf"] Dec 13 17:20:19 crc kubenswrapper[4989]: I1213 17:20:19.663020 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" event={"ID":"f3d71144-6e47-4a55-baa1-d523fff194e0","Type":"ContainerStarted","Data":"bafe1f947902afab206771051de55d32a81aba08c82c8f9b4954f3adb710a676"} Dec 13 17:20:19 crc kubenswrapper[4989]: I1213 17:20:19.663358 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:19 crc kubenswrapper[4989]: I1213 17:20:19.663380 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" event={"ID":"f3d71144-6e47-4a55-baa1-d523fff194e0","Type":"ContainerStarted","Data":"4cd9879590bda2d9b3ccd5052cdf1080c20eaeb226266a0d1bb2f6f0949592a0"} Dec 13 17:20:19 crc kubenswrapper[4989]: I1213 17:20:19.664408 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" event={"ID":"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4","Type":"ContainerStarted","Data":"1236d414f20704be59838ad980e9495c4d50138dc56ba630e40599c1938d911b"} Dec 13 17:20:19 crc kubenswrapper[4989]: I1213 17:20:19.664452 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" event={"ID":"be498ee8-a5fa-400a-b1e4-99e63b9eb6b4","Type":"ContainerStarted","Data":"111f29af66b3d35233bf54ea0a5e88b1bb455a3b95f2f30da2af958ff854fc01"} Dec 13 17:20:19 crc kubenswrapper[4989]: I1213 17:20:19.664826 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:19 crc kubenswrapper[4989]: I1213 17:20:19.668751 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" Dec 13 17:20:19 crc kubenswrapper[4989]: I1213 17:20:19.670081 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" Dec 13 17:20:19 crc kubenswrapper[4989]: I1213 17:20:19.682748 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-78ff4f79c4-lp55h" podStartSLOduration=3.682721804 podStartE2EDuration="3.682721804s" podCreationTimestamp="2025-12-13 17:20:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:20:19.681527696 +0000 UTC m=+234.287974864" watchObservedRunningTime="2025-12-13 17:20:19.682721804 +0000 UTC m=+234.289168942" Dec 13 17:20:19 crc kubenswrapper[4989]: I1213 17:20:19.709357 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5b6b7f6459-79dhf" podStartSLOduration=3.70934095 podStartE2EDuration="3.70934095s" podCreationTimestamp="2025-12-13 17:20:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:20:19.708206655 +0000 UTC m=+234.314653803" watchObservedRunningTime="2025-12-13 17:20:19.70934095 +0000 UTC m=+234.315788088" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.588544 4989 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.590167 4989 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.590314 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.590541 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f" gracePeriod=15 Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.590677 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e" gracePeriod=15 Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.590699 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14" gracePeriod=15 Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.590731 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986" gracePeriod=15 Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.590854 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c" gracePeriod=15 Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.591851 4989 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 13 17:20:24 crc kubenswrapper[4989]: E1213 17:20:24.593210 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593229 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 13 17:20:24 crc kubenswrapper[4989]: E1213 17:20:24.593246 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593254 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 13 17:20:24 crc kubenswrapper[4989]: E1213 17:20:24.593275 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593284 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 13 17:20:24 crc kubenswrapper[4989]: E1213 17:20:24.593296 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593302 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 17:20:24 crc kubenswrapper[4989]: E1213 17:20:24.593314 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593323 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 17:20:24 crc kubenswrapper[4989]: E1213 17:20:24.593332 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593341 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 13 17:20:24 crc kubenswrapper[4989]: E1213 17:20:24.593354 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593362 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593511 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593530 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593541 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593552 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593568 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.593576 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.629734 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.682188 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.682683 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.682730 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.682802 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.682823 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.682848 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.682866 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.682934 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.784050 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.784326 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.784471 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.784560 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.784633 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.784654 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.784398 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.784180 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.784805 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.784655 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.785066 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.785184 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.785189 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.785281 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.785368 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.785370 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: I1213 17:20:24.927177 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:20:24 crc kubenswrapper[4989]: W1213 17:20:24.948446 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-04d29d3283f0215677b423becb02f9caf28b0df2181df43f1357afbd1a3039b0 WatchSource:0}: Error finding container 04d29d3283f0215677b423becb02f9caf28b0df2181df43f1357afbd1a3039b0: Status 404 returned error can't find the container with id 04d29d3283f0215677b423becb02f9caf28b0df2181df43f1357afbd1a3039b0 Dec 13 17:20:24 crc kubenswrapper[4989]: E1213 17:20:24.951492 4989 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.47:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1880d611961fa6d1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-13 17:20:24.950679249 +0000 UTC m=+239.557126387,LastTimestamp:2025-12-13 17:20:24.950679249 +0000 UTC m=+239.557126387,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.079050 4989 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.079144 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.705300 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.707996 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.708967 4989 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986" exitCode=0 Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.709049 4989 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14" exitCode=0 Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.709127 4989 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e" exitCode=0 Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.709203 4989 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c" exitCode=2 Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.709051 4989 scope.go:117] "RemoveContainer" containerID="a3a817ae1cd33e3776735e917a5d611d43992f0c34fcba3b704e22cbe295e3db" Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.711115 4989 generic.go:334] "Generic (PLEG): container finished" podID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" containerID="a26c97b6dce23647e4f257ddecc35f4aacdea1ad61d4d915be34b68befe7b27a" exitCode=0 Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.711191 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"aa7cc4ba-4aff-402a-a2bd-9928394c24e5","Type":"ContainerDied","Data":"a26c97b6dce23647e4f257ddecc35f4aacdea1ad61d4d915be34b68befe7b27a"} Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.711940 4989 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.712457 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.712869 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.713368 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141"} Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.713395 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"04d29d3283f0215677b423becb02f9caf28b0df2181df43f1357afbd1a3039b0"} Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.714302 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.714745 4989 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:25 crc kubenswrapper[4989]: I1213 17:20:25.715141 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:26 crc kubenswrapper[4989]: I1213 17:20:26.026512 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:26 crc kubenswrapper[4989]: I1213 17:20:26.027546 4989 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:26 crc kubenswrapper[4989]: I1213 17:20:26.028276 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:26 crc kubenswrapper[4989]: I1213 17:20:26.751226 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 13 17:20:26 crc kubenswrapper[4989]: I1213 17:20:26.985252 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 13 17:20:26 crc kubenswrapper[4989]: I1213 17:20:26.986654 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:26 crc kubenswrapper[4989]: I1213 17:20:26.987608 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:26 crc kubenswrapper[4989]: I1213 17:20:26.988097 4989 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:26 crc kubenswrapper[4989]: I1213 17:20:26.988570 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.116550 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.116597 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.116699 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.116785 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.116924 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.117001 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.117752 4989 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.117776 4989 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.117809 4989 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.170722 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.171196 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.171394 4989 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.171861 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.218644 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-var-lock\") pod \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.218712 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kube-api-access\") pod \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.218756 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kubelet-dir\") pod \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\" (UID: \"aa7cc4ba-4aff-402a-a2bd-9928394c24e5\") " Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.218768 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-var-lock" (OuterVolumeSpecName: "var-lock") pod "aa7cc4ba-4aff-402a-a2bd-9928394c24e5" (UID: "aa7cc4ba-4aff-402a-a2bd-9928394c24e5"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.218895 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "aa7cc4ba-4aff-402a-a2bd-9928394c24e5" (UID: "aa7cc4ba-4aff-402a-a2bd-9928394c24e5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.219084 4989 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-var-lock\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.219102 4989 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.226746 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "aa7cc4ba-4aff-402a-a2bd-9928394c24e5" (UID: "aa7cc4ba-4aff-402a-a2bd-9928394c24e5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.320314 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa7cc4ba-4aff-402a-a2bd-9928394c24e5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.763254 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"aa7cc4ba-4aff-402a-a2bd-9928394c24e5","Type":"ContainerDied","Data":"de076427cc7c0505ad569a3acf5adb0327c6d19340eb5ba6606fc25d298a826a"} Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.763348 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de076427cc7c0505ad569a3acf5adb0327c6d19340eb5ba6606fc25d298a826a" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.763758 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.768069 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.770185 4989 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f" exitCode=0 Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.770288 4989 scope.go:117] "RemoveContainer" containerID="4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.770399 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.795638 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.796227 4989 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.797442 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.798714 4989 scope.go:117] "RemoveContainer" containerID="385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.799326 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.799553 4989 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.799837 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.817677 4989 scope.go:117] "RemoveContainer" containerID="dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.838362 4989 scope.go:117] "RemoveContainer" containerID="816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.855652 4989 scope.go:117] "RemoveContainer" containerID="da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.873285 4989 scope.go:117] "RemoveContainer" containerID="7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.899298 4989 scope.go:117] "RemoveContainer" containerID="4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986" Dec 13 17:20:27 crc kubenswrapper[4989]: E1213 17:20:27.900004 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\": container with ID starting with 4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986 not found: ID does not exist" containerID="4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.900059 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986"} err="failed to get container status \"4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\": rpc error: code = NotFound desc = could not find container \"4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986\": container with ID starting with 4becde2708057bac204f120910586a3ff71fa84a17cf93f80c474cc0e095f986 not found: ID does not exist" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.900089 4989 scope.go:117] "RemoveContainer" containerID="385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14" Dec 13 17:20:27 crc kubenswrapper[4989]: E1213 17:20:27.901204 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\": container with ID starting with 385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14 not found: ID does not exist" containerID="385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.901235 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14"} err="failed to get container status \"385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\": rpc error: code = NotFound desc = could not find container \"385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14\": container with ID starting with 385ad07d4eb60cb2240cf00a2382fa724f80f3b29c6edc351941d154fef32b14 not found: ID does not exist" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.901253 4989 scope.go:117] "RemoveContainer" containerID="dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e" Dec 13 17:20:27 crc kubenswrapper[4989]: E1213 17:20:27.901619 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\": container with ID starting with dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e not found: ID does not exist" containerID="dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.901687 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e"} err="failed to get container status \"dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\": rpc error: code = NotFound desc = could not find container \"dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e\": container with ID starting with dda51d1fb189f4a5cb9e230d96c1c86a023345868a93bc6359d80adf96ef852e not found: ID does not exist" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.901730 4989 scope.go:117] "RemoveContainer" containerID="816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c" Dec 13 17:20:27 crc kubenswrapper[4989]: E1213 17:20:27.902321 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\": container with ID starting with 816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c not found: ID does not exist" containerID="816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.902362 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c"} err="failed to get container status \"816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\": rpc error: code = NotFound desc = could not find container \"816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c\": container with ID starting with 816e687ea63f08874858a136ecd70e7369d0556fa6f8eae688da3062e9f6152c not found: ID does not exist" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.902390 4989 scope.go:117] "RemoveContainer" containerID="da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f" Dec 13 17:20:27 crc kubenswrapper[4989]: E1213 17:20:27.902705 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\": container with ID starting with da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f not found: ID does not exist" containerID="da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.902724 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f"} err="failed to get container status \"da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\": rpc error: code = NotFound desc = could not find container \"da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f\": container with ID starting with da71575b6c4240cd43182acfb437db46852d3f46cdc276cfd6891aacbc184b5f not found: ID does not exist" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.902758 4989 scope.go:117] "RemoveContainer" containerID="7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26" Dec 13 17:20:27 crc kubenswrapper[4989]: E1213 17:20:27.903161 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\": container with ID starting with 7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26 not found: ID does not exist" containerID="7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26" Dec 13 17:20:27 crc kubenswrapper[4989]: I1213 17:20:27.903192 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26"} err="failed to get container status \"7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\": rpc error: code = NotFound desc = could not find container \"7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26\": container with ID starting with 7c5ed6d7c40ccac399a82490e581fedc8592660ca0b3da6983c71e191e43db26 not found: ID does not exist" Dec 13 17:20:28 crc kubenswrapper[4989]: I1213 17:20:28.028295 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 13 17:20:30 crc kubenswrapper[4989]: E1213 17:20:30.930431 4989 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:30 crc kubenswrapper[4989]: E1213 17:20:30.933171 4989 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:30 crc kubenswrapper[4989]: E1213 17:20:30.933442 4989 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:30 crc kubenswrapper[4989]: E1213 17:20:30.933828 4989 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:30 crc kubenswrapper[4989]: E1213 17:20:30.934039 4989 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:30 crc kubenswrapper[4989]: I1213 17:20:30.934067 4989 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 13 17:20:30 crc kubenswrapper[4989]: E1213 17:20:30.934243 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="200ms" Dec 13 17:20:31 crc kubenswrapper[4989]: E1213 17:20:31.032832 4989 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.47:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" volumeName="registry-storage" Dec 13 17:20:31 crc kubenswrapper[4989]: E1213 17:20:31.135061 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="400ms" Dec 13 17:20:31 crc kubenswrapper[4989]: E1213 17:20:31.536062 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="800ms" Dec 13 17:20:32 crc kubenswrapper[4989]: E1213 17:20:32.337572 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="1.6s" Dec 13 17:20:33 crc kubenswrapper[4989]: E1213 17:20:33.939005 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="3.2s" Dec 13 17:20:34 crc kubenswrapper[4989]: E1213 17:20:34.101705 4989 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.47:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1880d611961fa6d1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-13 17:20:24.950679249 +0000 UTC m=+239.557126387,LastTimestamp:2025-12-13 17:20:24.950679249 +0000 UTC m=+239.557126387,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 13 17:20:36 crc kubenswrapper[4989]: I1213 17:20:36.017508 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:36 crc kubenswrapper[4989]: I1213 17:20:36.018273 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:36 crc kubenswrapper[4989]: I1213 17:20:36.640476 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" podUID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" containerName="oauth-openshift" containerID="cri-o://eeb229e7987e4ce1ac7bba37d265bacd5a8a9258a9033352e5fca9e9798c3e19" gracePeriod=15 Dec 13 17:20:36 crc kubenswrapper[4989]: I1213 17:20:36.831298 4989 generic.go:334] "Generic (PLEG): container finished" podID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" containerID="eeb229e7987e4ce1ac7bba37d265bacd5a8a9258a9033352e5fca9e9798c3e19" exitCode=0 Dec 13 17:20:36 crc kubenswrapper[4989]: I1213 17:20:36.831424 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" event={"ID":"badbe877-b1f8-40cf-9e7a-1d380a08f2bf","Type":"ContainerDied","Data":"eeb229e7987e4ce1ac7bba37d265bacd5a8a9258a9033352e5fca9e9798c3e19"} Dec 13 17:20:37 crc kubenswrapper[4989]: E1213 17:20:37.140229 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="6.4s" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.607412 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.607930 4989 status_manager.go:851] "Failed to get status for pod" podUID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dvtmv\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.608225 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.608660 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.649501 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-service-ca\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.649849 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-idp-0-file-data\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.649958 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-provider-selection\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650049 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-ocp-branding-template\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650124 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-error\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650239 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr8fr\" (UniqueName: \"kubernetes.io/projected/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-kube-api-access-rr8fr\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650362 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650382 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-login\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650471 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-serving-cert\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650511 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-dir\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650533 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-router-certs\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650560 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-trusted-ca-bundle\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650603 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-policies\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650628 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-session\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.650655 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-cliconfig\") pod \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\" (UID: \"badbe877-b1f8-40cf-9e7a-1d380a08f2bf\") " Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.651487 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.651654 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.653292 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.653555 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.656744 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.657144 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-kube-api-access-rr8fr" (OuterVolumeSpecName: "kube-api-access-rr8fr") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "kube-api-access-rr8fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.657758 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.661570 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.661770 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.664508 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.665531 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.668549 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.671017 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.672126 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "badbe877-b1f8-40cf-9e7a-1d380a08f2bf" (UID: "badbe877-b1f8-40cf-9e7a-1d380a08f2bf"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752607 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr8fr\" (UniqueName: \"kubernetes.io/projected/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-kube-api-access-rr8fr\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752667 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752683 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752699 4989 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752715 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752729 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752741 4989 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752766 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752778 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752807 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752820 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752833 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.752847 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/badbe877-b1f8-40cf-9e7a-1d380a08f2bf-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.838124 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" event={"ID":"badbe877-b1f8-40cf-9e7a-1d380a08f2bf","Type":"ContainerDied","Data":"c0d1984974d3afdd74517ec1ebcfc303dae284feff3e7ff2a4a35bc78bff2296"} Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.838179 4989 scope.go:117] "RemoveContainer" containerID="eeb229e7987e4ce1ac7bba37d265bacd5a8a9258a9033352e5fca9e9798c3e19" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.838210 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.839041 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.839399 4989 status_manager.go:851] "Failed to get status for pod" podUID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dvtmv\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.839635 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.854097 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.854421 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:37 crc kubenswrapper[4989]: I1213 17:20:37.855077 4989 status_manager.go:851] "Failed to get status for pod" podUID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dvtmv\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.013973 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.015441 4989 status_manager.go:851] "Failed to get status for pod" podUID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dvtmv\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.015966 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.017188 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.029988 4989 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.030021 4989 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:20:38 crc kubenswrapper[4989]: E1213 17:20:38.030383 4989 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.030954 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:38 crc kubenswrapper[4989]: W1213 17:20:38.051750 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-3e016ed860741f0cd3b07ad757b6e96624b23cbd56ffea12b8fcb6cd2ae9288b WatchSource:0}: Error finding container 3e016ed860741f0cd3b07ad757b6e96624b23cbd56ffea12b8fcb6cd2ae9288b: Status 404 returned error can't find the container with id 3e016ed860741f0cd3b07ad757b6e96624b23cbd56ffea12b8fcb6cd2ae9288b Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.847316 4989 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="9a53360213338c306be592c0172dd5eaf7f15c4a8f64ac919d527ed21aae2a31" exitCode=0 Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.847402 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"9a53360213338c306be592c0172dd5eaf7f15c4a8f64ac919d527ed21aae2a31"} Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.849015 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3e016ed860741f0cd3b07ad757b6e96624b23cbd56ffea12b8fcb6cd2ae9288b"} Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.849697 4989 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.849742 4989 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:20:38 crc kubenswrapper[4989]: E1213 17:20:38.850331 4989 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.850387 4989 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.851134 4989 status_manager.go:851] "Failed to get status for pod" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:38 crc kubenswrapper[4989]: I1213 17:20:38.851817 4989 status_manager.go:851] "Failed to get status for pod" podUID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" pod="openshift-authentication/oauth-openshift-558db77b4-dvtmv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-dvtmv\": dial tcp 38.102.83.47:6443: connect: connection refused" Dec 13 17:20:39 crc kubenswrapper[4989]: I1213 17:20:39.876733 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 13 17:20:39 crc kubenswrapper[4989]: I1213 17:20:39.877016 4989 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d" exitCode=1 Dec 13 17:20:39 crc kubenswrapper[4989]: I1213 17:20:39.877073 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d"} Dec 13 17:20:39 crc kubenswrapper[4989]: I1213 17:20:39.877524 4989 scope.go:117] "RemoveContainer" containerID="a2ebe62634e9d6d8f385dd8b65b76c601491fb6e72542b144cc341491cef9f6d" Dec 13 17:20:39 crc kubenswrapper[4989]: I1213 17:20:39.884761 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bb2916c340556d1eaf7a761894d8c3de943f53b1208e21fa3fde9e6df1fb91ce"} Dec 13 17:20:39 crc kubenswrapper[4989]: I1213 17:20:39.884830 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f2fc8a2d739817a31cb33baeabc8bc48171b27af67bf25090a2a6838cf419592"} Dec 13 17:20:39 crc kubenswrapper[4989]: I1213 17:20:39.884841 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9716d57577d7fb0d2ed590ba940e75b77a48c7273558cc9c9b670c58d6c779eb"} Dec 13 17:20:39 crc kubenswrapper[4989]: I1213 17:20:39.884850 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8528d8805736c4396f18d858acb5df2d44fa8d18da534206d7feba0432f35edf"} Dec 13 17:20:40 crc kubenswrapper[4989]: I1213 17:20:40.902891 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e9a789483677fd94b18b431f06542a3243a5c65b6de5807cd61dcb49051c454d"} Dec 13 17:20:40 crc kubenswrapper[4989]: I1213 17:20:40.903208 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:40 crc kubenswrapper[4989]: I1213 17:20:40.902997 4989 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:20:40 crc kubenswrapper[4989]: I1213 17:20:40.903266 4989 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:20:40 crc kubenswrapper[4989]: I1213 17:20:40.906277 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 13 17:20:40 crc kubenswrapper[4989]: I1213 17:20:40.906319 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6a454bc220e0f88f25d4c815ec057a7c60fa9be01781be03b3faa22562f6b1aa"} Dec 13 17:20:43 crc kubenswrapper[4989]: I1213 17:20:43.031546 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:43 crc kubenswrapper[4989]: I1213 17:20:43.031625 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:43 crc kubenswrapper[4989]: I1213 17:20:43.039200 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:45 crc kubenswrapper[4989]: I1213 17:20:45.917717 4989 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:45 crc kubenswrapper[4989]: I1213 17:20:45.938575 4989 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:20:45 crc kubenswrapper[4989]: I1213 17:20:45.938605 4989 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:20:45 crc kubenswrapper[4989]: I1213 17:20:45.942389 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:20:46 crc kubenswrapper[4989]: I1213 17:20:46.021360 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:20:46 crc kubenswrapper[4989]: I1213 17:20:46.027414 4989 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="407f4626-24aa-42ee-b62d-1eed8ae4a94c" Dec 13 17:20:46 crc kubenswrapper[4989]: I1213 17:20:46.943775 4989 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:20:46 crc kubenswrapper[4989]: I1213 17:20:46.944311 4989 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:20:46 crc kubenswrapper[4989]: I1213 17:20:46.946756 4989 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="407f4626-24aa-42ee-b62d-1eed8ae4a94c" Dec 13 17:20:47 crc kubenswrapper[4989]: I1213 17:20:47.049082 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:20:47 crc kubenswrapper[4989]: I1213 17:20:47.054749 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:20:55 crc kubenswrapper[4989]: I1213 17:20:55.518294 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 13 17:20:55 crc kubenswrapper[4989]: I1213 17:20:55.823379 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 13 17:20:55 crc kubenswrapper[4989]: I1213 17:20:55.980372 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 13 17:20:56 crc kubenswrapper[4989]: I1213 17:20:56.023473 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 13 17:20:56 crc kubenswrapper[4989]: I1213 17:20:56.185810 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 13 17:20:56 crc kubenswrapper[4989]: I1213 17:20:56.260296 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 13 17:20:56 crc kubenswrapper[4989]: I1213 17:20:56.692404 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 13 17:20:56 crc kubenswrapper[4989]: I1213 17:20:56.887995 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 13 17:20:56 crc kubenswrapper[4989]: I1213 17:20:56.916212 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 13 17:20:57 crc kubenswrapper[4989]: I1213 17:20:57.013328 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 13 17:20:57 crc kubenswrapper[4989]: I1213 17:20:57.109138 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 13 17:20:57 crc kubenswrapper[4989]: I1213 17:20:57.202870 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 13 17:20:57 crc kubenswrapper[4989]: I1213 17:20:57.624084 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 13 17:20:57 crc kubenswrapper[4989]: I1213 17:20:57.987990 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 13 17:20:57 crc kubenswrapper[4989]: I1213 17:20:57.990151 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.133197 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.180515 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.181489 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.220307 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.237438 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.287633 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.412083 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.457444 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.461038 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.644331 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.725712 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 13 17:20:58 crc kubenswrapper[4989]: I1213 17:20:58.802733 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.079507 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.111634 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.160695 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.178667 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.215369 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.324877 4989 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.335866 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.337942 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.344672 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.458138 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.583942 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.632438 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.817267 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.835464 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 13 17:20:59 crc kubenswrapper[4989]: I1213 17:20:59.847218 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.032950 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.164923 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.264598 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.271342 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.306758 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.499238 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.606509 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.607679 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.662688 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.684117 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.813813 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 13 17:21:00 crc kubenswrapper[4989]: I1213 17:21:00.906577 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.001232 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.074726 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.078439 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.100636 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.157974 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.207488 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.253716 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.306575 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.429534 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.436859 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.515444 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.521379 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.526072 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.546479 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.580708 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.638575 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.794285 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.807658 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.820320 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.843291 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.866456 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.916747 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.934444 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 13 17:21:01 crc kubenswrapper[4989]: I1213 17:21:01.959018 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.066953 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.173313 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.218160 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.465812 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.512355 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.521688 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.521751 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.522861 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.700623 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.740421 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.769403 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.818653 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.835541 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.858709 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.968868 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 13 17:21:02 crc kubenswrapper[4989]: I1213 17:21:02.979657 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.093392 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.181073 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.520727 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.560099 4989 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.560862 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.562462 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=39.562445884 podStartE2EDuration="39.562445884s" podCreationTimestamp="2025-12-13 17:20:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:20:45.517530263 +0000 UTC m=+260.123977401" watchObservedRunningTime="2025-12-13 17:21:03.562445884 +0000 UTC m=+278.168893012" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.565584 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-dvtmv","openshift-kube-apiserver/kube-apiserver-crc"] Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.565632 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v"] Dec 13 17:21:03 crc kubenswrapper[4989]: E1213 17:21:03.565857 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" containerName="installer" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.565873 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" containerName="installer" Dec 13 17:21:03 crc kubenswrapper[4989]: E1213 17:21:03.565889 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" containerName="oauth-openshift" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.565896 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" containerName="oauth-openshift" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.566025 4989 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.566042 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa7cc4ba-4aff-402a-a2bd-9928394c24e5" containerName="installer" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.566048 4989 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0d12847a-6be0-4ab1-8052-cd417f1525c9" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.566053 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" containerName="oauth-openshift" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.566438 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.567617 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.571105 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.571232 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.572578 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.574069 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.574553 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.574597 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.574967 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.575404 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.575574 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.575668 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.575760 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.578531 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.582083 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.587751 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.587882 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.608862 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.608835284 podStartE2EDuration="18.608835284s" podCreationTimestamp="2025-12-13 17:20:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:21:03.607158258 +0000 UTC m=+278.213605406" watchObservedRunningTime="2025-12-13 17:21:03.608835284 +0000 UTC m=+278.215282442" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.611169 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.654323 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.663361 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.669437 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.718650 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dklhq\" (UniqueName: \"kubernetes.io/projected/6714bf3f-565c-49e4-be55-e049e504d600-kube-api-access-dklhq\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.718992 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-session\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.719124 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.719236 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.719344 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.719442 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.719546 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.719657 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.719776 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.719920 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.720044 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.720197 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-audit-policies\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.720382 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6714bf3f-565c-49e4-be55-e049e504d600-audit-dir\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.720493 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.785603 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.794927 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821372 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-audit-policies\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821423 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6714bf3f-565c-49e4-be55-e049e504d600-audit-dir\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821447 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821466 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dklhq\" (UniqueName: \"kubernetes.io/projected/6714bf3f-565c-49e4-be55-e049e504d600-kube-api-access-dklhq\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821488 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-session\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821504 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821520 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821541 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821557 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821576 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821603 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821624 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821645 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.821662 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.822522 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6714bf3f-565c-49e4-be55-e049e504d600-audit-dir\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.822933 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.823038 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-audit-policies\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.823170 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.823902 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.827282 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.827708 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.829043 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.829115 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.829175 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.829609 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-system-session\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.835826 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.835867 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6714bf3f-565c-49e4-be55-e049e504d600-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.838882 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dklhq\" (UniqueName: \"kubernetes.io/projected/6714bf3f-565c-49e4-be55-e049e504d600-kube-api-access-dklhq\") pod \"oauth-openshift-7d9bbcf4d4-5wc2v\" (UID: \"6714bf3f-565c-49e4-be55-e049e504d600\") " pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.891385 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.962715 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.969326 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 13 17:21:03 crc kubenswrapper[4989]: I1213 17:21:03.988628 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.021254 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="badbe877-b1f8-40cf-9e7a-1d380a08f2bf" path="/var/lib/kubelet/pods/badbe877-b1f8-40cf-9e7a-1d380a08f2bf/volumes" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.027473 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.036513 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.066250 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.086028 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.228670 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.245278 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.343251 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.624318 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.762571 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.776201 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.778431 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.889357 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.908091 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v"] Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.917654 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.934329 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.974736 4989 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 13 17:21:04 crc kubenswrapper[4989]: I1213 17:21:04.990074 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.000285 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.024315 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.024623 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.024975 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.083557 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.125585 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.197428 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.260257 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.308530 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.338964 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.357998 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v"] Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.393193 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.494248 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.518708 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.560169 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.693393 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.716664 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.817466 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.853031 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.885925 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.887903 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 13 17:21:05 crc kubenswrapper[4989]: I1213 17:21:05.938052 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.057446 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.070634 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" event={"ID":"6714bf3f-565c-49e4-be55-e049e504d600","Type":"ContainerStarted","Data":"2206fde55071ee9eb52951f12b9780f0f83f4226f6d2fb4551be1e29be36e927"} Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.070681 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" event={"ID":"6714bf3f-565c-49e4-be55-e049e504d600","Type":"ContainerStarted","Data":"eaef1430b021f9c742c2330830b055a6e8d2cfab042e55acb2d01102da007eaa"} Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.071827 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.092057 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.095076 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" podStartSLOduration=55.095057526 podStartE2EDuration="55.095057526s" podCreationTimestamp="2025-12-13 17:20:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:21:06.094164856 +0000 UTC m=+280.700612004" watchObservedRunningTime="2025-12-13 17:21:06.095057526 +0000 UTC m=+280.701504674" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.098355 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.179965 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.219102 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7d9bbcf4d4-5wc2v" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.222123 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.278258 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.401705 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.433191 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.542067 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.557008 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.593642 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.651406 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.688595 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.846157 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.849140 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.866482 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.897249 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.898406 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.901182 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 13 17:21:06 crc kubenswrapper[4989]: I1213 17:21:06.970019 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.034855 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.050882 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.058634 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.085728 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.091864 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.270195 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.281677 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.347185 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.347288 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.429158 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.438372 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.448897 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.481967 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.498600 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.619730 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.641723 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.643467 4989 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.686470 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.800917 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.812349 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 13 17:21:07 crc kubenswrapper[4989]: I1213 17:21:07.973386 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.008949 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.129590 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.203105 4989 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.203587 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141" gracePeriod=5 Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.242700 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.247936 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.375452 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.400369 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.443851 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.539369 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.606513 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.685657 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.733150 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.775000 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.890191 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 13 17:21:08 crc kubenswrapper[4989]: I1213 17:21:08.925737 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.020105 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.183469 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.245080 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.344059 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.519589 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.553746 4989 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.558908 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.683651 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.723603 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.734349 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.740668 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.837929 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.850157 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.893287 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 13 17:21:09 crc kubenswrapper[4989]: I1213 17:21:09.922138 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 13 17:21:10 crc kubenswrapper[4989]: I1213 17:21:10.086069 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 13 17:21:10 crc kubenswrapper[4989]: I1213 17:21:10.089642 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 13 17:21:10 crc kubenswrapper[4989]: I1213 17:21:10.110193 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 13 17:21:10 crc kubenswrapper[4989]: I1213 17:21:10.194581 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 13 17:21:10 crc kubenswrapper[4989]: I1213 17:21:10.242466 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 13 17:21:10 crc kubenswrapper[4989]: I1213 17:21:10.356129 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 13 17:21:10 crc kubenswrapper[4989]: I1213 17:21:10.367186 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 13 17:21:10 crc kubenswrapper[4989]: I1213 17:21:10.441436 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 13 17:21:10 crc kubenswrapper[4989]: I1213 17:21:10.456723 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 13 17:21:10 crc kubenswrapper[4989]: I1213 17:21:10.557089 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 13 17:21:10 crc kubenswrapper[4989]: I1213 17:21:10.596025 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 13 17:21:11 crc kubenswrapper[4989]: I1213 17:21:11.005399 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 13 17:21:11 crc kubenswrapper[4989]: I1213 17:21:11.225220 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 13 17:21:11 crc kubenswrapper[4989]: I1213 17:21:11.305647 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 13 17:21:11 crc kubenswrapper[4989]: I1213 17:21:11.523270 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 13 17:21:11 crc kubenswrapper[4989]: I1213 17:21:11.819664 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 13 17:21:11 crc kubenswrapper[4989]: I1213 17:21:11.934283 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 13 17:21:12 crc kubenswrapper[4989]: I1213 17:21:12.000708 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 13 17:21:12 crc kubenswrapper[4989]: I1213 17:21:12.064131 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 13 17:21:12 crc kubenswrapper[4989]: I1213 17:21:12.386884 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 13 17:21:12 crc kubenswrapper[4989]: I1213 17:21:12.445808 4989 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 13 17:21:12 crc kubenswrapper[4989]: I1213 17:21:12.956603 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.792466 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.792544 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.859943 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.859983 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.860066 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.860072 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.860120 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.860131 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.860151 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.860154 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.860179 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.860338 4989 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.860349 4989 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.860357 4989 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.860365 4989 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.867258 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:21:13 crc kubenswrapper[4989]: I1213 17:21:13.961292 4989 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.024738 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.025028 4989 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.035131 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.035166 4989 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="5eb20fdd-843a-4c94-899a-a26528aa62a8" Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.038409 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.038449 4989 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="5eb20fdd-843a-4c94-899a-a26528aa62a8" Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.112450 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.112509 4989 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141" exitCode=137 Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.112561 4989 scope.go:117] "RemoveContainer" containerID="604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141" Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.112574 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.129266 4989 scope.go:117] "RemoveContainer" containerID="604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141" Dec 13 17:21:14 crc kubenswrapper[4989]: E1213 17:21:14.129607 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141\": container with ID starting with 604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141 not found: ID does not exist" containerID="604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141" Dec 13 17:21:14 crc kubenswrapper[4989]: I1213 17:21:14.129647 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141"} err="failed to get container status \"604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141\": rpc error: code = NotFound desc = could not find container \"604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141\": container with ID starting with 604107226d2632abd74cfca474b3aea013c49da3b1688578aa54ed453415b141 not found: ID does not exist" Dec 13 17:21:25 crc kubenswrapper[4989]: I1213 17:21:25.862726 4989 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 13 17:21:52 crc kubenswrapper[4989]: I1213 17:21:52.395754 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 13 17:21:52 crc kubenswrapper[4989]: I1213 17:21:52.983144 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cgvr5"] Dec 13 17:21:52 crc kubenswrapper[4989]: E1213 17:21:52.983494 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 13 17:21:52 crc kubenswrapper[4989]: I1213 17:21:52.983521 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 13 17:21:52 crc kubenswrapper[4989]: I1213 17:21:52.983706 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 13 17:21:52 crc kubenswrapper[4989]: I1213 17:21:52.984359 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.010541 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cgvr5"] Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.105664 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67125893-3009-4c78-b5db-863a56460e28-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.105734 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.105777 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67125893-3009-4c78-b5db-863a56460e28-registry-tls\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.105882 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67125893-3009-4c78-b5db-863a56460e28-trusted-ca\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.105949 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlqxt\" (UniqueName: \"kubernetes.io/projected/67125893-3009-4c78-b5db-863a56460e28-kube-api-access-hlqxt\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.105977 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67125893-3009-4c78-b5db-863a56460e28-registry-certificates\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.106304 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67125893-3009-4c78-b5db-863a56460e28-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.106365 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67125893-3009-4c78-b5db-863a56460e28-bound-sa-token\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.136364 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.207914 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67125893-3009-4c78-b5db-863a56460e28-registry-tls\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.207972 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67125893-3009-4c78-b5db-863a56460e28-trusted-ca\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.207993 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlqxt\" (UniqueName: \"kubernetes.io/projected/67125893-3009-4c78-b5db-863a56460e28-kube-api-access-hlqxt\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.208013 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67125893-3009-4c78-b5db-863a56460e28-registry-certificates\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.208047 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67125893-3009-4c78-b5db-863a56460e28-bound-sa-token\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.208068 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67125893-3009-4c78-b5db-863a56460e28-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.208104 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67125893-3009-4c78-b5db-863a56460e28-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.208647 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67125893-3009-4c78-b5db-863a56460e28-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.209320 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67125893-3009-4c78-b5db-863a56460e28-registry-certificates\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.210178 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67125893-3009-4c78-b5db-863a56460e28-trusted-ca\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.215102 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67125893-3009-4c78-b5db-863a56460e28-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.215140 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67125893-3009-4c78-b5db-863a56460e28-registry-tls\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.226893 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlqxt\" (UniqueName: \"kubernetes.io/projected/67125893-3009-4c78-b5db-863a56460e28-kube-api-access-hlqxt\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.231467 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67125893-3009-4c78-b5db-863a56460e28-bound-sa-token\") pod \"image-registry-66df7c8f76-cgvr5\" (UID: \"67125893-3009-4c78-b5db-863a56460e28\") " pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.318284 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:53 crc kubenswrapper[4989]: I1213 17:21:53.713285 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cgvr5"] Dec 13 17:21:54 crc kubenswrapper[4989]: I1213 17:21:54.350097 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" event={"ID":"67125893-3009-4c78-b5db-863a56460e28","Type":"ContainerStarted","Data":"a1ddbb51399117db5cfefc0d24bd1f9894765f5b579c87668d26fefc399b5ca2"} Dec 13 17:21:54 crc kubenswrapper[4989]: I1213 17:21:54.350324 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" event={"ID":"67125893-3009-4c78-b5db-863a56460e28","Type":"ContainerStarted","Data":"febb4d1e50f36eb6308a2bf8124d1c351695df1e3da9c6e49d92c10f211084be"} Dec 13 17:21:54 crc kubenswrapper[4989]: I1213 17:21:54.350347 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:21:54 crc kubenswrapper[4989]: I1213 17:21:54.370343 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" podStartSLOduration=2.370323489 podStartE2EDuration="2.370323489s" podCreationTimestamp="2025-12-13 17:21:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:21:54.369040536 +0000 UTC m=+328.975487674" watchObservedRunningTime="2025-12-13 17:21:54.370323489 +0000 UTC m=+328.976770637" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.248404 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tbwsb"] Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.251243 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tbwsb" podUID="5fda1871-6df5-4256-8a90-74bb43a13290" containerName="registry-server" containerID="cri-o://a1656b0632dec392795049dd397c6d85cad331fd22ddd95abd97d473b043de57" gracePeriod=30 Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.262328 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d8k9w"] Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.262732 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d8k9w" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerName="registry-server" containerID="cri-o://ab85b2383289d5502ef3478c71aa9db066983438101a3eae96a5fed3e2a920ae" gracePeriod=30 Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.276888 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k5w28"] Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.278950 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" podUID="bdf2d561-45dc-4672-929e-df9803b1d581" containerName="marketplace-operator" containerID="cri-o://10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94" gracePeriod=30 Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.290037 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6f77"] Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.290403 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x6f77" podUID="00527548-b884-4adf-8c65-4c6ca4367d77" containerName="registry-server" containerID="cri-o://93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6" gracePeriod=30 Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.301843 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j5xss"] Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.302232 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j5xss" podUID="37118126-08b8-46bd-bde4-a4faace357a2" containerName="registry-server" containerID="cri-o://9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3" gracePeriod=30 Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.305043 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wl78k"] Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.309206 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.309728 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wl78k"] Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.371916 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a56dab99-6f88-4da4-af60-913e0b4c3be6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wl78k\" (UID: \"a56dab99-6f88-4da4-af60-913e0b4c3be6\") " pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.372032 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkzrv\" (UniqueName: \"kubernetes.io/projected/a56dab99-6f88-4da4-af60-913e0b4c3be6-kube-api-access-bkzrv\") pod \"marketplace-operator-79b997595-wl78k\" (UID: \"a56dab99-6f88-4da4-af60-913e0b4c3be6\") " pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.372108 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a56dab99-6f88-4da4-af60-913e0b4c3be6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wl78k\" (UID: \"a56dab99-6f88-4da4-af60-913e0b4c3be6\") " pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.427532 4989 generic.go:334] "Generic (PLEG): container finished" podID="5fda1871-6df5-4256-8a90-74bb43a13290" containerID="a1656b0632dec392795049dd397c6d85cad331fd22ddd95abd97d473b043de57" exitCode=0 Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.427605 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tbwsb" event={"ID":"5fda1871-6df5-4256-8a90-74bb43a13290","Type":"ContainerDied","Data":"a1656b0632dec392795049dd397c6d85cad331fd22ddd95abd97d473b043de57"} Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.436247 4989 generic.go:334] "Generic (PLEG): container finished" podID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerID="ab85b2383289d5502ef3478c71aa9db066983438101a3eae96a5fed3e2a920ae" exitCode=0 Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.436288 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8k9w" event={"ID":"199e44f9-fad2-4498-8c5b-685a10c97ab8","Type":"ContainerDied","Data":"ab85b2383289d5502ef3478c71aa9db066983438101a3eae96a5fed3e2a920ae"} Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.481655 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a56dab99-6f88-4da4-af60-913e0b4c3be6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wl78k\" (UID: \"a56dab99-6f88-4da4-af60-913e0b4c3be6\") " pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.481725 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkzrv\" (UniqueName: \"kubernetes.io/projected/a56dab99-6f88-4da4-af60-913e0b4c3be6-kube-api-access-bkzrv\") pod \"marketplace-operator-79b997595-wl78k\" (UID: \"a56dab99-6f88-4da4-af60-913e0b4c3be6\") " pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.481761 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a56dab99-6f88-4da4-af60-913e0b4c3be6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wl78k\" (UID: \"a56dab99-6f88-4da4-af60-913e0b4c3be6\") " pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.483434 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a56dab99-6f88-4da4-af60-913e0b4c3be6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wl78k\" (UID: \"a56dab99-6f88-4da4-af60-913e0b4c3be6\") " pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.517567 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a56dab99-6f88-4da4-af60-913e0b4c3be6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wl78k\" (UID: \"a56dab99-6f88-4da4-af60-913e0b4c3be6\") " pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.530686 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkzrv\" (UniqueName: \"kubernetes.io/projected/a56dab99-6f88-4da4-af60-913e0b4c3be6-kube-api-access-bkzrv\") pod \"marketplace-operator-79b997595-wl78k\" (UID: \"a56dab99-6f88-4da4-af60-913e0b4c3be6\") " pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.634362 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.782695 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.844951 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.848712 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.869487 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.890227 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-catalog-content\") pod \"5fda1871-6df5-4256-8a90-74bb43a13290\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.890359 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-utilities\") pod \"5fda1871-6df5-4256-8a90-74bb43a13290\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.890411 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtv88\" (UniqueName: \"kubernetes.io/projected/5fda1871-6df5-4256-8a90-74bb43a13290-kube-api-access-qtv88\") pod \"5fda1871-6df5-4256-8a90-74bb43a13290\" (UID: \"5fda1871-6df5-4256-8a90-74bb43a13290\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.894070 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-utilities" (OuterVolumeSpecName: "utilities") pod "5fda1871-6df5-4256-8a90-74bb43a13290" (UID: "5fda1871-6df5-4256-8a90-74bb43a13290"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.894840 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fda1871-6df5-4256-8a90-74bb43a13290-kube-api-access-qtv88" (OuterVolumeSpecName: "kube-api-access-qtv88") pod "5fda1871-6df5-4256-8a90-74bb43a13290" (UID: "5fda1871-6df5-4256-8a90-74bb43a13290"). InnerVolumeSpecName "kube-api-access-qtv88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.952815 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.962038 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fda1871-6df5-4256-8a90-74bb43a13290" (UID: "5fda1871-6df5-4256-8a90-74bb43a13290"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.991862 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8nvm\" (UniqueName: \"kubernetes.io/projected/199e44f9-fad2-4498-8c5b-685a10c97ab8-kube-api-access-w8nvm\") pod \"199e44f9-fad2-4498-8c5b-685a10c97ab8\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.991941 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-catalog-content\") pod \"37118126-08b8-46bd-bde4-a4faace357a2\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.992012 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-catalog-content\") pod \"199e44f9-fad2-4498-8c5b-685a10c97ab8\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.992069 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-utilities\") pod \"00527548-b884-4adf-8c65-4c6ca4367d77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.992098 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmj9m\" (UniqueName: \"kubernetes.io/projected/00527548-b884-4adf-8c65-4c6ca4367d77-kube-api-access-vmj9m\") pod \"00527548-b884-4adf-8c65-4c6ca4367d77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.992155 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-utilities\") pod \"37118126-08b8-46bd-bde4-a4faace357a2\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.992218 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-utilities\") pod \"199e44f9-fad2-4498-8c5b-685a10c97ab8\" (UID: \"199e44f9-fad2-4498-8c5b-685a10c97ab8\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.992244 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-catalog-content\") pod \"00527548-b884-4adf-8c65-4c6ca4367d77\" (UID: \"00527548-b884-4adf-8c65-4c6ca4367d77\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.992275 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdrlp\" (UniqueName: \"kubernetes.io/projected/37118126-08b8-46bd-bde4-a4faace357a2-kube-api-access-zdrlp\") pod \"37118126-08b8-46bd-bde4-a4faace357a2\" (UID: \"37118126-08b8-46bd-bde4-a4faace357a2\") " Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.992583 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.992600 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fda1871-6df5-4256-8a90-74bb43a13290-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.992613 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtv88\" (UniqueName: \"kubernetes.io/projected/5fda1871-6df5-4256-8a90-74bb43a13290-kube-api-access-qtv88\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.993547 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-utilities" (OuterVolumeSpecName: "utilities") pod "00527548-b884-4adf-8c65-4c6ca4367d77" (UID: "00527548-b884-4adf-8c65-4c6ca4367d77"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.993709 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-utilities" (OuterVolumeSpecName: "utilities") pod "37118126-08b8-46bd-bde4-a4faace357a2" (UID: "37118126-08b8-46bd-bde4-a4faace357a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.994320 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-utilities" (OuterVolumeSpecName: "utilities") pod "199e44f9-fad2-4498-8c5b-685a10c97ab8" (UID: "199e44f9-fad2-4498-8c5b-685a10c97ab8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.995425 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00527548-b884-4adf-8c65-4c6ca4367d77-kube-api-access-vmj9m" (OuterVolumeSpecName: "kube-api-access-vmj9m") pod "00527548-b884-4adf-8c65-4c6ca4367d77" (UID: "00527548-b884-4adf-8c65-4c6ca4367d77"). InnerVolumeSpecName "kube-api-access-vmj9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.995535 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/199e44f9-fad2-4498-8c5b-685a10c97ab8-kube-api-access-w8nvm" (OuterVolumeSpecName: "kube-api-access-w8nvm") pod "199e44f9-fad2-4498-8c5b-685a10c97ab8" (UID: "199e44f9-fad2-4498-8c5b-685a10c97ab8"). InnerVolumeSpecName "kube-api-access-w8nvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:22:04 crc kubenswrapper[4989]: I1213 17:22:04.995912 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37118126-08b8-46bd-bde4-a4faace357a2-kube-api-access-zdrlp" (OuterVolumeSpecName: "kube-api-access-zdrlp") pod "37118126-08b8-46bd-bde4-a4faace357a2" (UID: "37118126-08b8-46bd-bde4-a4faace357a2"). InnerVolumeSpecName "kube-api-access-zdrlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.014289 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00527548-b884-4adf-8c65-4c6ca4367d77" (UID: "00527548-b884-4adf-8c65-4c6ca4367d77"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.049687 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "199e44f9-fad2-4498-8c5b-685a10c97ab8" (UID: "199e44f9-fad2-4498-8c5b-685a10c97ab8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.093631 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-operator-metrics\") pod \"bdf2d561-45dc-4672-929e-df9803b1d581\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.093688 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-trusted-ca\") pod \"bdf2d561-45dc-4672-929e-df9803b1d581\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.093908 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtd24\" (UniqueName: \"kubernetes.io/projected/bdf2d561-45dc-4672-929e-df9803b1d581-kube-api-access-wtd24\") pod \"bdf2d561-45dc-4672-929e-df9803b1d581\" (UID: \"bdf2d561-45dc-4672-929e-df9803b1d581\") " Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.094243 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8nvm\" (UniqueName: \"kubernetes.io/projected/199e44f9-fad2-4498-8c5b-685a10c97ab8-kube-api-access-w8nvm\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.094388 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.094523 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.094557 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmj9m\" (UniqueName: \"kubernetes.io/projected/00527548-b884-4adf-8c65-4c6ca4367d77-kube-api-access-vmj9m\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.094573 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.094583 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/199e44f9-fad2-4498-8c5b-685a10c97ab8-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.094593 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00527548-b884-4adf-8c65-4c6ca4367d77-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.094605 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdrlp\" (UniqueName: \"kubernetes.io/projected/37118126-08b8-46bd-bde4-a4faace357a2-kube-api-access-zdrlp\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.095419 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "bdf2d561-45dc-4672-929e-df9803b1d581" (UID: "bdf2d561-45dc-4672-929e-df9803b1d581"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.096649 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "bdf2d561-45dc-4672-929e-df9803b1d581" (UID: "bdf2d561-45dc-4672-929e-df9803b1d581"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.097141 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdf2d561-45dc-4672-929e-df9803b1d581-kube-api-access-wtd24" (OuterVolumeSpecName: "kube-api-access-wtd24") pod "bdf2d561-45dc-4672-929e-df9803b1d581" (UID: "bdf2d561-45dc-4672-929e-df9803b1d581"). InnerVolumeSpecName "kube-api-access-wtd24". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.102955 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37118126-08b8-46bd-bde4-a4faace357a2" (UID: "37118126-08b8-46bd-bde4-a4faace357a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.155025 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wl78k"] Dec 13 17:22:05 crc kubenswrapper[4989]: W1213 17:22:05.157938 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda56dab99_6f88_4da4_af60_913e0b4c3be6.slice/crio-54b201f8f2e3d4d479369cfe6a0ee0a3ca02168d0158c382da00fc09c1c45216 WatchSource:0}: Error finding container 54b201f8f2e3d4d479369cfe6a0ee0a3ca02168d0158c382da00fc09c1c45216: Status 404 returned error can't find the container with id 54b201f8f2e3d4d479369cfe6a0ee0a3ca02168d0158c382da00fc09c1c45216 Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.195469 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtd24\" (UniqueName: \"kubernetes.io/projected/bdf2d561-45dc-4672-929e-df9803b1d581-kube-api-access-wtd24\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.195496 4989 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.195508 4989 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bdf2d561-45dc-4672-929e-df9803b1d581-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.195517 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37118126-08b8-46bd-bde4-a4faace357a2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.444061 4989 generic.go:334] "Generic (PLEG): container finished" podID="bdf2d561-45dc-4672-929e-df9803b1d581" containerID="10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94" exitCode=0 Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.444105 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" event={"ID":"bdf2d561-45dc-4672-929e-df9803b1d581","Type":"ContainerDied","Data":"10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94"} Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.444450 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" event={"ID":"bdf2d561-45dc-4672-929e-df9803b1d581","Type":"ContainerDied","Data":"6666f9a0b3c2f434ce57129a75e4b5e616a86120c0b23216c93e6c06c7bcb422"} Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.444161 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k5w28" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.444479 4989 scope.go:117] "RemoveContainer" containerID="10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.446387 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" event={"ID":"a56dab99-6f88-4da4-af60-913e0b4c3be6","Type":"ContainerStarted","Data":"ce46662ec92833fb85dcf0933ee56773cb57884649e5500b0884e1a94d9fac51"} Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.446421 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" event={"ID":"a56dab99-6f88-4da4-af60-913e0b4c3be6","Type":"ContainerStarted","Data":"54b201f8f2e3d4d479369cfe6a0ee0a3ca02168d0158c382da00fc09c1c45216"} Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.447233 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.448971 4989 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wl78k container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.66:8080/healthz\": dial tcp 10.217.0.66:8080: connect: connection refused" start-of-body= Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.449020 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" podUID="a56dab99-6f88-4da4-af60-913e0b4c3be6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.66:8080/healthz\": dial tcp 10.217.0.66:8080: connect: connection refused" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.450722 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8k9w" event={"ID":"199e44f9-fad2-4498-8c5b-685a10c97ab8","Type":"ContainerDied","Data":"347203dc3f1b8e5d74c7205e16fa93ebb4acb2705f257180850b5881aefed598"} Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.450843 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d8k9w" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.454824 4989 generic.go:334] "Generic (PLEG): container finished" podID="00527548-b884-4adf-8c65-4c6ca4367d77" containerID="93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6" exitCode=0 Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.454900 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6f77" event={"ID":"00527548-b884-4adf-8c65-4c6ca4367d77","Type":"ContainerDied","Data":"93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6"} Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.454929 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6f77" event={"ID":"00527548-b884-4adf-8c65-4c6ca4367d77","Type":"ContainerDied","Data":"dfa1ac17780b89489726deadd00d46d6b0522c9988333eb88be4dfaee66b39d5"} Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.454989 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6f77" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.465613 4989 scope.go:117] "RemoveContainer" containerID="10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94" Dec 13 17:22:05 crc kubenswrapper[4989]: E1213 17:22:05.467600 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94\": container with ID starting with 10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94 not found: ID does not exist" containerID="10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.467702 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94"} err="failed to get container status \"10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94\": rpc error: code = NotFound desc = could not find container \"10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94\": container with ID starting with 10a2a6c5fa1548840fa03d662d7eaed4c2122c3c430481e1d5093ae95cc24b94 not found: ID does not exist" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.467807 4989 scope.go:117] "RemoveContainer" containerID="ab85b2383289d5502ef3478c71aa9db066983438101a3eae96a5fed3e2a920ae" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.473362 4989 generic.go:334] "Generic (PLEG): container finished" podID="37118126-08b8-46bd-bde4-a4faace357a2" containerID="9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3" exitCode=0 Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.473452 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5xss" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.473481 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5xss" event={"ID":"37118126-08b8-46bd-bde4-a4faace357a2","Type":"ContainerDied","Data":"9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3"} Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.473723 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5xss" event={"ID":"37118126-08b8-46bd-bde4-a4faace357a2","Type":"ContainerDied","Data":"21d5c11dd6ceed384e4462b7097e56d2b5d4f66c1c756b20c476dc7575fd81ba"} Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.479121 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" podStartSLOduration=1.4790920889999999 podStartE2EDuration="1.479092089s" podCreationTimestamp="2025-12-13 17:22:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:22:05.474081872 +0000 UTC m=+340.080529020" watchObservedRunningTime="2025-12-13 17:22:05.479092089 +0000 UTC m=+340.085539237" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.480059 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tbwsb" event={"ID":"5fda1871-6df5-4256-8a90-74bb43a13290","Type":"ContainerDied","Data":"0754b9172aa9d21d604466e8af3d73f4cf7ba11627ec142a5b46e446393c403e"} Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.480149 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tbwsb" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.491537 4989 scope.go:117] "RemoveContainer" containerID="e34f6ae15f2817dd660f08ff87ffa0f459f213a30867bbc2529c825bf1ab463e" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.509614 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k5w28"] Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.515085 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k5w28"] Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.527665 4989 scope.go:117] "RemoveContainer" containerID="cb57bfd12d0420a459b71aa75c063c99bd1b5b6bcb393c579f15a0f8f451d3fc" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.527889 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6f77"] Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.533748 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6f77"] Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.540203 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d8k9w"] Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.546991 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d8k9w"] Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.553403 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tbwsb"] Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.554045 4989 scope.go:117] "RemoveContainer" containerID="93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.559934 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tbwsb"] Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.563876 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j5xss"] Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.566929 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j5xss"] Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.575083 4989 scope.go:117] "RemoveContainer" containerID="ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.587961 4989 scope.go:117] "RemoveContainer" containerID="df86ff0474ef1e17bd90f9237fd5a38c3c3f55e819bb38d37e2067b5ef0d518e" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.615742 4989 scope.go:117] "RemoveContainer" containerID="93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6" Dec 13 17:22:05 crc kubenswrapper[4989]: E1213 17:22:05.616316 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6\": container with ID starting with 93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6 not found: ID does not exist" containerID="93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.616367 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6"} err="failed to get container status \"93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6\": rpc error: code = NotFound desc = could not find container \"93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6\": container with ID starting with 93f5b161125c636908d4a6438632ac883972a37895d06ecdf4d08c15f113efc6 not found: ID does not exist" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.616406 4989 scope.go:117] "RemoveContainer" containerID="ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb" Dec 13 17:22:05 crc kubenswrapper[4989]: E1213 17:22:05.620340 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb\": container with ID starting with ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb not found: ID does not exist" containerID="ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.620374 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb"} err="failed to get container status \"ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb\": rpc error: code = NotFound desc = could not find container \"ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb\": container with ID starting with ec87c9b340a45e4c93075aee72c3559ad9b7590a7dcbf0efd79dbb5d508302eb not found: ID does not exist" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.620433 4989 scope.go:117] "RemoveContainer" containerID="df86ff0474ef1e17bd90f9237fd5a38c3c3f55e819bb38d37e2067b5ef0d518e" Dec 13 17:22:05 crc kubenswrapper[4989]: E1213 17:22:05.620814 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df86ff0474ef1e17bd90f9237fd5a38c3c3f55e819bb38d37e2067b5ef0d518e\": container with ID starting with df86ff0474ef1e17bd90f9237fd5a38c3c3f55e819bb38d37e2067b5ef0d518e not found: ID does not exist" containerID="df86ff0474ef1e17bd90f9237fd5a38c3c3f55e819bb38d37e2067b5ef0d518e" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.620850 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df86ff0474ef1e17bd90f9237fd5a38c3c3f55e819bb38d37e2067b5ef0d518e"} err="failed to get container status \"df86ff0474ef1e17bd90f9237fd5a38c3c3f55e819bb38d37e2067b5ef0d518e\": rpc error: code = NotFound desc = could not find container \"df86ff0474ef1e17bd90f9237fd5a38c3c3f55e819bb38d37e2067b5ef0d518e\": container with ID starting with df86ff0474ef1e17bd90f9237fd5a38c3c3f55e819bb38d37e2067b5ef0d518e not found: ID does not exist" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.620873 4989 scope.go:117] "RemoveContainer" containerID="9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.633996 4989 scope.go:117] "RemoveContainer" containerID="d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.648576 4989 scope.go:117] "RemoveContainer" containerID="d35400cb1c4a5dbdd7d518e731cffebf5ed10600f07df3e07ebbecd133d3fff9" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.663160 4989 scope.go:117] "RemoveContainer" containerID="9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3" Dec 13 17:22:05 crc kubenswrapper[4989]: E1213 17:22:05.663673 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3\": container with ID starting with 9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3 not found: ID does not exist" containerID="9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.663993 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3"} err="failed to get container status \"9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3\": rpc error: code = NotFound desc = could not find container \"9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3\": container with ID starting with 9f9be6562ed4a7fd4686329ec1523cab7ace0ca0d2df6b51283806fc1cd4ace3 not found: ID does not exist" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.664020 4989 scope.go:117] "RemoveContainer" containerID="d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9" Dec 13 17:22:05 crc kubenswrapper[4989]: E1213 17:22:05.664386 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9\": container with ID starting with d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9 not found: ID does not exist" containerID="d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.664407 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9"} err="failed to get container status \"d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9\": rpc error: code = NotFound desc = could not find container \"d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9\": container with ID starting with d5f456e863cc4ccbbce46af0768c454faa6d585780fd8f3649a195e3810343f9 not found: ID does not exist" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.664423 4989 scope.go:117] "RemoveContainer" containerID="d35400cb1c4a5dbdd7d518e731cffebf5ed10600f07df3e07ebbecd133d3fff9" Dec 13 17:22:05 crc kubenswrapper[4989]: E1213 17:22:05.664710 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d35400cb1c4a5dbdd7d518e731cffebf5ed10600f07df3e07ebbecd133d3fff9\": container with ID starting with d35400cb1c4a5dbdd7d518e731cffebf5ed10600f07df3e07ebbecd133d3fff9 not found: ID does not exist" containerID="d35400cb1c4a5dbdd7d518e731cffebf5ed10600f07df3e07ebbecd133d3fff9" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.664732 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d35400cb1c4a5dbdd7d518e731cffebf5ed10600f07df3e07ebbecd133d3fff9"} err="failed to get container status \"d35400cb1c4a5dbdd7d518e731cffebf5ed10600f07df3e07ebbecd133d3fff9\": rpc error: code = NotFound desc = could not find container \"d35400cb1c4a5dbdd7d518e731cffebf5ed10600f07df3e07ebbecd133d3fff9\": container with ID starting with d35400cb1c4a5dbdd7d518e731cffebf5ed10600f07df3e07ebbecd133d3fff9 not found: ID does not exist" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.664746 4989 scope.go:117] "RemoveContainer" containerID="a1656b0632dec392795049dd397c6d85cad331fd22ddd95abd97d473b043de57" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.680095 4989 scope.go:117] "RemoveContainer" containerID="94c5c65f53f5a6981040392429390ffcd9447873d9779d4d86cd4ef43985c586" Dec 13 17:22:05 crc kubenswrapper[4989]: I1213 17:22:05.698869 4989 scope.go:117] "RemoveContainer" containerID="8a3da68912b07caabef9849d5882773774bc75c66ae98f10845555017a583045" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.026231 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00527548-b884-4adf-8c65-4c6ca4367d77" path="/var/lib/kubelet/pods/00527548-b884-4adf-8c65-4c6ca4367d77/volumes" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.027635 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" path="/var/lib/kubelet/pods/199e44f9-fad2-4498-8c5b-685a10c97ab8/volumes" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.028849 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37118126-08b8-46bd-bde4-a4faace357a2" path="/var/lib/kubelet/pods/37118126-08b8-46bd-bde4-a4faace357a2/volumes" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.030591 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fda1871-6df5-4256-8a90-74bb43a13290" path="/var/lib/kubelet/pods/5fda1871-6df5-4256-8a90-74bb43a13290/volumes" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.031247 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdf2d561-45dc-4672-929e-df9803b1d581" path="/var/lib/kubelet/pods/bdf2d561-45dc-4672-929e-df9803b1d581/volumes" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.461828 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g542s"] Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462055 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00527548-b884-4adf-8c65-4c6ca4367d77" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462067 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="00527548-b884-4adf-8c65-4c6ca4367d77" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462078 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerName="extract-content" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462085 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerName="extract-content" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462094 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fda1871-6df5-4256-8a90-74bb43a13290" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462100 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fda1871-6df5-4256-8a90-74bb43a13290" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462109 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerName="extract-utilities" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462115 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerName="extract-utilities" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462122 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf2d561-45dc-4672-929e-df9803b1d581" containerName="marketplace-operator" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462129 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf2d561-45dc-4672-929e-df9803b1d581" containerName="marketplace-operator" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462139 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fda1871-6df5-4256-8a90-74bb43a13290" containerName="extract-utilities" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462144 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fda1871-6df5-4256-8a90-74bb43a13290" containerName="extract-utilities" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462152 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37118126-08b8-46bd-bde4-a4faace357a2" containerName="extract-content" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462158 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="37118126-08b8-46bd-bde4-a4faace357a2" containerName="extract-content" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462168 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37118126-08b8-46bd-bde4-a4faace357a2" containerName="extract-utilities" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462175 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="37118126-08b8-46bd-bde4-a4faace357a2" containerName="extract-utilities" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462183 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37118126-08b8-46bd-bde4-a4faace357a2" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462188 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="37118126-08b8-46bd-bde4-a4faace357a2" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462198 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00527548-b884-4adf-8c65-4c6ca4367d77" containerName="extract-utilities" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462203 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="00527548-b884-4adf-8c65-4c6ca4367d77" containerName="extract-utilities" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462213 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00527548-b884-4adf-8c65-4c6ca4367d77" containerName="extract-content" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462219 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="00527548-b884-4adf-8c65-4c6ca4367d77" containerName="extract-content" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462225 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fda1871-6df5-4256-8a90-74bb43a13290" containerName="extract-content" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462231 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fda1871-6df5-4256-8a90-74bb43a13290" containerName="extract-content" Dec 13 17:22:06 crc kubenswrapper[4989]: E1213 17:22:06.462237 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462242 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462341 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf2d561-45dc-4672-929e-df9803b1d581" containerName="marketplace-operator" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462352 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fda1871-6df5-4256-8a90-74bb43a13290" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462362 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="37118126-08b8-46bd-bde4-a4faace357a2" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462371 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="199e44f9-fad2-4498-8c5b-685a10c97ab8" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.462379 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="00527548-b884-4adf-8c65-4c6ca4367d77" containerName="registry-server" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.463212 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.466648 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.471121 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g542s"] Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.494810 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wl78k" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.628588 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98203275-48dd-4bcd-a4c0-709142c9e024-utilities\") pod \"redhat-marketplace-g542s\" (UID: \"98203275-48dd-4bcd-a4c0-709142c9e024\") " pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.628697 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98203275-48dd-4bcd-a4c0-709142c9e024-catalog-content\") pod \"redhat-marketplace-g542s\" (UID: \"98203275-48dd-4bcd-a4c0-709142c9e024\") " pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.628756 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpbcd\" (UniqueName: \"kubernetes.io/projected/98203275-48dd-4bcd-a4c0-709142c9e024-kube-api-access-dpbcd\") pod \"redhat-marketplace-g542s\" (UID: \"98203275-48dd-4bcd-a4c0-709142c9e024\") " pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.653218 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r59g9"] Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.654858 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.658554 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.663872 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r59g9"] Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.730432 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc88ed3d-34b8-4968-a381-9bf45fe374c1-catalog-content\") pod \"certified-operators-r59g9\" (UID: \"bc88ed3d-34b8-4968-a381-9bf45fe374c1\") " pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.730479 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc88ed3d-34b8-4968-a381-9bf45fe374c1-utilities\") pod \"certified-operators-r59g9\" (UID: \"bc88ed3d-34b8-4968-a381-9bf45fe374c1\") " pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.730510 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98203275-48dd-4bcd-a4c0-709142c9e024-catalog-content\") pod \"redhat-marketplace-g542s\" (UID: \"98203275-48dd-4bcd-a4c0-709142c9e024\") " pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.730539 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc9gq\" (UniqueName: \"kubernetes.io/projected/bc88ed3d-34b8-4968-a381-9bf45fe374c1-kube-api-access-dc9gq\") pod \"certified-operators-r59g9\" (UID: \"bc88ed3d-34b8-4968-a381-9bf45fe374c1\") " pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.730745 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpbcd\" (UniqueName: \"kubernetes.io/projected/98203275-48dd-4bcd-a4c0-709142c9e024-kube-api-access-dpbcd\") pod \"redhat-marketplace-g542s\" (UID: \"98203275-48dd-4bcd-a4c0-709142c9e024\") " pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.730937 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98203275-48dd-4bcd-a4c0-709142c9e024-catalog-content\") pod \"redhat-marketplace-g542s\" (UID: \"98203275-48dd-4bcd-a4c0-709142c9e024\") " pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.730937 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98203275-48dd-4bcd-a4c0-709142c9e024-utilities\") pod \"redhat-marketplace-g542s\" (UID: \"98203275-48dd-4bcd-a4c0-709142c9e024\") " pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.731490 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98203275-48dd-4bcd-a4c0-709142c9e024-utilities\") pod \"redhat-marketplace-g542s\" (UID: \"98203275-48dd-4bcd-a4c0-709142c9e024\") " pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.757186 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpbcd\" (UniqueName: \"kubernetes.io/projected/98203275-48dd-4bcd-a4c0-709142c9e024-kube-api-access-dpbcd\") pod \"redhat-marketplace-g542s\" (UID: \"98203275-48dd-4bcd-a4c0-709142c9e024\") " pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.789093 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.832651 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc88ed3d-34b8-4968-a381-9bf45fe374c1-catalog-content\") pod \"certified-operators-r59g9\" (UID: \"bc88ed3d-34b8-4968-a381-9bf45fe374c1\") " pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.832902 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc88ed3d-34b8-4968-a381-9bf45fe374c1-utilities\") pod \"certified-operators-r59g9\" (UID: \"bc88ed3d-34b8-4968-a381-9bf45fe374c1\") " pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.832938 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc9gq\" (UniqueName: \"kubernetes.io/projected/bc88ed3d-34b8-4968-a381-9bf45fe374c1-kube-api-access-dc9gq\") pod \"certified-operators-r59g9\" (UID: \"bc88ed3d-34b8-4968-a381-9bf45fe374c1\") " pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.833244 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc88ed3d-34b8-4968-a381-9bf45fe374c1-catalog-content\") pod \"certified-operators-r59g9\" (UID: \"bc88ed3d-34b8-4968-a381-9bf45fe374c1\") " pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.833254 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc88ed3d-34b8-4968-a381-9bf45fe374c1-utilities\") pod \"certified-operators-r59g9\" (UID: \"bc88ed3d-34b8-4968-a381-9bf45fe374c1\") " pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:06 crc kubenswrapper[4989]: I1213 17:22:06.848522 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc9gq\" (UniqueName: \"kubernetes.io/projected/bc88ed3d-34b8-4968-a381-9bf45fe374c1-kube-api-access-dc9gq\") pod \"certified-operators-r59g9\" (UID: \"bc88ed3d-34b8-4968-a381-9bf45fe374c1\") " pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:07 crc kubenswrapper[4989]: I1213 17:22:07.008043 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:07 crc kubenswrapper[4989]: I1213 17:22:07.180243 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g542s"] Dec 13 17:22:07 crc kubenswrapper[4989]: W1213 17:22:07.182332 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98203275_48dd_4bcd_a4c0_709142c9e024.slice/crio-689f62d853a81d5ef0ec44dddf5d32dd3f6b5529b575c953a9cc8854764a6083 WatchSource:0}: Error finding container 689f62d853a81d5ef0ec44dddf5d32dd3f6b5529b575c953a9cc8854764a6083: Status 404 returned error can't find the container with id 689f62d853a81d5ef0ec44dddf5d32dd3f6b5529b575c953a9cc8854764a6083 Dec 13 17:22:07 crc kubenswrapper[4989]: I1213 17:22:07.370753 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r59g9"] Dec 13 17:22:07 crc kubenswrapper[4989]: I1213 17:22:07.499520 4989 generic.go:334] "Generic (PLEG): container finished" podID="98203275-48dd-4bcd-a4c0-709142c9e024" containerID="6040e4158eca35bfd4a41528854f26afa972a6d16d050c1654373f6ceb1cd19f" exitCode=0 Dec 13 17:22:07 crc kubenswrapper[4989]: I1213 17:22:07.499579 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g542s" event={"ID":"98203275-48dd-4bcd-a4c0-709142c9e024","Type":"ContainerDied","Data":"6040e4158eca35bfd4a41528854f26afa972a6d16d050c1654373f6ceb1cd19f"} Dec 13 17:22:07 crc kubenswrapper[4989]: I1213 17:22:07.499609 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g542s" event={"ID":"98203275-48dd-4bcd-a4c0-709142c9e024","Type":"ContainerStarted","Data":"689f62d853a81d5ef0ec44dddf5d32dd3f6b5529b575c953a9cc8854764a6083"} Dec 13 17:22:07 crc kubenswrapper[4989]: I1213 17:22:07.502222 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r59g9" event={"ID":"bc88ed3d-34b8-4968-a381-9bf45fe374c1","Type":"ContainerStarted","Data":"655ddc0507111c21df96b049f8592f9a4c0e9f2b97925136446b0da47d3dcf65"} Dec 13 17:22:07 crc kubenswrapper[4989]: I1213 17:22:07.502275 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r59g9" event={"ID":"bc88ed3d-34b8-4968-a381-9bf45fe374c1","Type":"ContainerStarted","Data":"8091bf8529c35c111cf979b622973b191bfd8677ff5a66a0098617a2af937778"} Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.509039 4989 generic.go:334] "Generic (PLEG): container finished" podID="bc88ed3d-34b8-4968-a381-9bf45fe374c1" containerID="655ddc0507111c21df96b049f8592f9a4c0e9f2b97925136446b0da47d3dcf65" exitCode=0 Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.509086 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r59g9" event={"ID":"bc88ed3d-34b8-4968-a381-9bf45fe374c1","Type":"ContainerDied","Data":"655ddc0507111c21df96b049f8592f9a4c0e9f2b97925136446b0da47d3dcf65"} Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.509393 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r59g9" event={"ID":"bc88ed3d-34b8-4968-a381-9bf45fe374c1","Type":"ContainerStarted","Data":"41ebe6a922f3c8b115730b696feb194328c502324fa08269a6fd58100ce4952f"} Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.513942 4989 generic.go:334] "Generic (PLEG): container finished" podID="98203275-48dd-4bcd-a4c0-709142c9e024" containerID="76ea013bdcb88d6711f86d9e95728bfb622f77490dead0f74ca934a563ba7277" exitCode=0 Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.514059 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g542s" event={"ID":"98203275-48dd-4bcd-a4c0-709142c9e024","Type":"ContainerDied","Data":"76ea013bdcb88d6711f86d9e95728bfb622f77490dead0f74ca934a563ba7277"} Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.860599 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xsmj4"] Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.861545 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.864968 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.875825 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xsmj4"] Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.960703 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4wtb\" (UniqueName: \"kubernetes.io/projected/1e515c25-23f4-49f2-bc3f-0fbb5e8046be-kube-api-access-r4wtb\") pod \"redhat-operators-xsmj4\" (UID: \"1e515c25-23f4-49f2-bc3f-0fbb5e8046be\") " pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.961227 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e515c25-23f4-49f2-bc3f-0fbb5e8046be-catalog-content\") pod \"redhat-operators-xsmj4\" (UID: \"1e515c25-23f4-49f2-bc3f-0fbb5e8046be\") " pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:08 crc kubenswrapper[4989]: I1213 17:22:08.961295 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e515c25-23f4-49f2-bc3f-0fbb5e8046be-utilities\") pod \"redhat-operators-xsmj4\" (UID: \"1e515c25-23f4-49f2-bc3f-0fbb5e8046be\") " pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.063592 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6bhwp"] Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.065843 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.066564 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4wtb\" (UniqueName: \"kubernetes.io/projected/1e515c25-23f4-49f2-bc3f-0fbb5e8046be-kube-api-access-r4wtb\") pod \"redhat-operators-xsmj4\" (UID: \"1e515c25-23f4-49f2-bc3f-0fbb5e8046be\") " pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.067244 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e515c25-23f4-49f2-bc3f-0fbb5e8046be-catalog-content\") pod \"redhat-operators-xsmj4\" (UID: \"1e515c25-23f4-49f2-bc3f-0fbb5e8046be\") " pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.067341 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e515c25-23f4-49f2-bc3f-0fbb5e8046be-utilities\") pod \"redhat-operators-xsmj4\" (UID: \"1e515c25-23f4-49f2-bc3f-0fbb5e8046be\") " pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.067685 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e515c25-23f4-49f2-bc3f-0fbb5e8046be-catalog-content\") pod \"redhat-operators-xsmj4\" (UID: \"1e515c25-23f4-49f2-bc3f-0fbb5e8046be\") " pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.068010 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.072041 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e515c25-23f4-49f2-bc3f-0fbb5e8046be-utilities\") pod \"redhat-operators-xsmj4\" (UID: \"1e515c25-23f4-49f2-bc3f-0fbb5e8046be\") " pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.074518 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6bhwp"] Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.092848 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4wtb\" (UniqueName: \"kubernetes.io/projected/1e515c25-23f4-49f2-bc3f-0fbb5e8046be-kube-api-access-r4wtb\") pod \"redhat-operators-xsmj4\" (UID: \"1e515c25-23f4-49f2-bc3f-0fbb5e8046be\") " pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.169018 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/981677fb-9915-45ba-83c3-bb602abaf7d4-catalog-content\") pod \"community-operators-6bhwp\" (UID: \"981677fb-9915-45ba-83c3-bb602abaf7d4\") " pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.169096 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h2r7\" (UniqueName: \"kubernetes.io/projected/981677fb-9915-45ba-83c3-bb602abaf7d4-kube-api-access-6h2r7\") pod \"community-operators-6bhwp\" (UID: \"981677fb-9915-45ba-83c3-bb602abaf7d4\") " pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.169165 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/981677fb-9915-45ba-83c3-bb602abaf7d4-utilities\") pod \"community-operators-6bhwp\" (UID: \"981677fb-9915-45ba-83c3-bb602abaf7d4\") " pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.191601 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.271027 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/981677fb-9915-45ba-83c3-bb602abaf7d4-catalog-content\") pod \"community-operators-6bhwp\" (UID: \"981677fb-9915-45ba-83c3-bb602abaf7d4\") " pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.271305 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h2r7\" (UniqueName: \"kubernetes.io/projected/981677fb-9915-45ba-83c3-bb602abaf7d4-kube-api-access-6h2r7\") pod \"community-operators-6bhwp\" (UID: \"981677fb-9915-45ba-83c3-bb602abaf7d4\") " pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.271360 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/981677fb-9915-45ba-83c3-bb602abaf7d4-utilities\") pod \"community-operators-6bhwp\" (UID: \"981677fb-9915-45ba-83c3-bb602abaf7d4\") " pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.271997 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/981677fb-9915-45ba-83c3-bb602abaf7d4-utilities\") pod \"community-operators-6bhwp\" (UID: \"981677fb-9915-45ba-83c3-bb602abaf7d4\") " pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.272162 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/981677fb-9915-45ba-83c3-bb602abaf7d4-catalog-content\") pod \"community-operators-6bhwp\" (UID: \"981677fb-9915-45ba-83c3-bb602abaf7d4\") " pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.294410 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h2r7\" (UniqueName: \"kubernetes.io/projected/981677fb-9915-45ba-83c3-bb602abaf7d4-kube-api-access-6h2r7\") pod \"community-operators-6bhwp\" (UID: \"981677fb-9915-45ba-83c3-bb602abaf7d4\") " pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.384252 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.520079 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g542s" event={"ID":"98203275-48dd-4bcd-a4c0-709142c9e024","Type":"ContainerStarted","Data":"ea1268218cf3bf2911d4d8e151bbd914f464552a44c2c2655362ded6efb9fd50"} Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.522637 4989 generic.go:334] "Generic (PLEG): container finished" podID="bc88ed3d-34b8-4968-a381-9bf45fe374c1" containerID="41ebe6a922f3c8b115730b696feb194328c502324fa08269a6fd58100ce4952f" exitCode=0 Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.522679 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r59g9" event={"ID":"bc88ed3d-34b8-4968-a381-9bf45fe374c1","Type":"ContainerDied","Data":"41ebe6a922f3c8b115730b696feb194328c502324fa08269a6fd58100ce4952f"} Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.536659 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g542s" podStartSLOduration=1.9153976259999999 podStartE2EDuration="3.536644106s" podCreationTimestamp="2025-12-13 17:22:06 +0000 UTC" firstStartedPulling="2025-12-13 17:22:07.500870915 +0000 UTC m=+342.107318053" lastFinishedPulling="2025-12-13 17:22:09.122117385 +0000 UTC m=+343.728564533" observedRunningTime="2025-12-13 17:22:09.535879172 +0000 UTC m=+344.142326320" watchObservedRunningTime="2025-12-13 17:22:09.536644106 +0000 UTC m=+344.143091244" Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.596376 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xsmj4"] Dec 13 17:22:09 crc kubenswrapper[4989]: I1213 17:22:09.802354 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6bhwp"] Dec 13 17:22:10 crc kubenswrapper[4989]: I1213 17:22:10.530022 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r59g9" event={"ID":"bc88ed3d-34b8-4968-a381-9bf45fe374c1","Type":"ContainerStarted","Data":"445cadadbfa3a5e93fa388ffbcd9b54c5ae45b5c7f82bb71bddb19788ef21e66"} Dec 13 17:22:10 crc kubenswrapper[4989]: I1213 17:22:10.533252 4989 generic.go:334] "Generic (PLEG): container finished" podID="981677fb-9915-45ba-83c3-bb602abaf7d4" containerID="8f14babfd73f219033b7c4b271ff0bed4fbe8722111866c348f8cb116d05dcd7" exitCode=0 Dec 13 17:22:10 crc kubenswrapper[4989]: I1213 17:22:10.533310 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bhwp" event={"ID":"981677fb-9915-45ba-83c3-bb602abaf7d4","Type":"ContainerDied","Data":"8f14babfd73f219033b7c4b271ff0bed4fbe8722111866c348f8cb116d05dcd7"} Dec 13 17:22:10 crc kubenswrapper[4989]: I1213 17:22:10.533327 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bhwp" event={"ID":"981677fb-9915-45ba-83c3-bb602abaf7d4","Type":"ContainerStarted","Data":"a15a8cb620ce40153d801998adc3bb6a49e23c97e3f417e99acce228d9e1a698"} Dec 13 17:22:10 crc kubenswrapper[4989]: I1213 17:22:10.534877 4989 generic.go:334] "Generic (PLEG): container finished" podID="1e515c25-23f4-49f2-bc3f-0fbb5e8046be" containerID="1edc6b8b28649d9525d0b857f47f3f1d53035e9227fc03eecd335f5f6e70b391" exitCode=0 Dec 13 17:22:10 crc kubenswrapper[4989]: I1213 17:22:10.534902 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsmj4" event={"ID":"1e515c25-23f4-49f2-bc3f-0fbb5e8046be","Type":"ContainerDied","Data":"1edc6b8b28649d9525d0b857f47f3f1d53035e9227fc03eecd335f5f6e70b391"} Dec 13 17:22:10 crc kubenswrapper[4989]: I1213 17:22:10.534930 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsmj4" event={"ID":"1e515c25-23f4-49f2-bc3f-0fbb5e8046be","Type":"ContainerStarted","Data":"8b923e1e1cf3cca72c096274a7494fe06156a125e06898d0b010e1c2d028a40f"} Dec 13 17:22:10 crc kubenswrapper[4989]: I1213 17:22:10.548073 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r59g9" podStartSLOduration=2.052563383 podStartE2EDuration="4.548051787s" podCreationTimestamp="2025-12-13 17:22:06 +0000 UTC" firstStartedPulling="2025-12-13 17:22:07.503643757 +0000 UTC m=+342.110090895" lastFinishedPulling="2025-12-13 17:22:09.999132161 +0000 UTC m=+344.605579299" observedRunningTime="2025-12-13 17:22:10.545865275 +0000 UTC m=+345.152312413" watchObservedRunningTime="2025-12-13 17:22:10.548051787 +0000 UTC m=+345.154498925" Dec 13 17:22:11 crc kubenswrapper[4989]: I1213 17:22:11.541995 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsmj4" event={"ID":"1e515c25-23f4-49f2-bc3f-0fbb5e8046be","Type":"ContainerStarted","Data":"8619863be1d0f9e16dd6f04c731d4525ad7635dc3b2203fcfcc0595d6026816b"} Dec 13 17:22:11 crc kubenswrapper[4989]: I1213 17:22:11.544415 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bhwp" event={"ID":"981677fb-9915-45ba-83c3-bb602abaf7d4","Type":"ContainerStarted","Data":"92517a6eb89f9d5a111fe54e3ed0ced73dab10ced17459a7e144eb856fc1f166"} Dec 13 17:22:12 crc kubenswrapper[4989]: I1213 17:22:12.550685 4989 generic.go:334] "Generic (PLEG): container finished" podID="981677fb-9915-45ba-83c3-bb602abaf7d4" containerID="92517a6eb89f9d5a111fe54e3ed0ced73dab10ced17459a7e144eb856fc1f166" exitCode=0 Dec 13 17:22:12 crc kubenswrapper[4989]: I1213 17:22:12.550743 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bhwp" event={"ID":"981677fb-9915-45ba-83c3-bb602abaf7d4","Type":"ContainerDied","Data":"92517a6eb89f9d5a111fe54e3ed0ced73dab10ced17459a7e144eb856fc1f166"} Dec 13 17:22:12 crc kubenswrapper[4989]: I1213 17:22:12.554834 4989 generic.go:334] "Generic (PLEG): container finished" podID="1e515c25-23f4-49f2-bc3f-0fbb5e8046be" containerID="8619863be1d0f9e16dd6f04c731d4525ad7635dc3b2203fcfcc0595d6026816b" exitCode=0 Dec 13 17:22:12 crc kubenswrapper[4989]: I1213 17:22:12.554931 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsmj4" event={"ID":"1e515c25-23f4-49f2-bc3f-0fbb5e8046be","Type":"ContainerDied","Data":"8619863be1d0f9e16dd6f04c731d4525ad7635dc3b2203fcfcc0595d6026816b"} Dec 13 17:22:13 crc kubenswrapper[4989]: I1213 17:22:13.326486 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-cgvr5" Dec 13 17:22:13 crc kubenswrapper[4989]: I1213 17:22:13.381194 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9blxx"] Dec 13 17:22:14 crc kubenswrapper[4989]: I1213 17:22:14.568637 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bhwp" event={"ID":"981677fb-9915-45ba-83c3-bb602abaf7d4","Type":"ContainerStarted","Data":"eb4fac6413003c60497ecddab1f6f0377fbc043f567a4c2a794c362505bffcb4"} Dec 13 17:22:14 crc kubenswrapper[4989]: I1213 17:22:14.571138 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xsmj4" event={"ID":"1e515c25-23f4-49f2-bc3f-0fbb5e8046be","Type":"ContainerStarted","Data":"ac548c2343dd9d49c19ecc9b389ec40cb89381f23e966ce173b12339076dd233"} Dec 13 17:22:14 crc kubenswrapper[4989]: I1213 17:22:14.591201 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6bhwp" podStartSLOduration=3.11882323 podStartE2EDuration="5.591179426s" podCreationTimestamp="2025-12-13 17:22:09 +0000 UTC" firstStartedPulling="2025-12-13 17:22:10.535655705 +0000 UTC m=+345.142102843" lastFinishedPulling="2025-12-13 17:22:13.008011901 +0000 UTC m=+347.614459039" observedRunningTime="2025-12-13 17:22:14.585925542 +0000 UTC m=+349.192372690" watchObservedRunningTime="2025-12-13 17:22:14.591179426 +0000 UTC m=+349.197626564" Dec 13 17:22:14 crc kubenswrapper[4989]: I1213 17:22:14.617546 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xsmj4" podStartSLOduration=4.1552038079999996 podStartE2EDuration="6.617523871s" podCreationTimestamp="2025-12-13 17:22:08 +0000 UTC" firstStartedPulling="2025-12-13 17:22:10.536559975 +0000 UTC m=+345.143007103" lastFinishedPulling="2025-12-13 17:22:12.998880018 +0000 UTC m=+347.605327166" observedRunningTime="2025-12-13 17:22:14.61326633 +0000 UTC m=+349.219713478" watchObservedRunningTime="2025-12-13 17:22:14.617523871 +0000 UTC m=+349.223971009" Dec 13 17:22:16 crc kubenswrapper[4989]: I1213 17:22:16.789656 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:16 crc kubenswrapper[4989]: I1213 17:22:16.791886 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:16 crc kubenswrapper[4989]: I1213 17:22:16.834429 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:17 crc kubenswrapper[4989]: I1213 17:22:17.008585 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:17 crc kubenswrapper[4989]: I1213 17:22:17.008648 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:17 crc kubenswrapper[4989]: I1213 17:22:17.046390 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:17 crc kubenswrapper[4989]: I1213 17:22:17.641955 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g542s" Dec 13 17:22:17 crc kubenswrapper[4989]: I1213 17:22:17.653462 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r59g9" Dec 13 17:22:19 crc kubenswrapper[4989]: I1213 17:22:19.192007 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:19 crc kubenswrapper[4989]: I1213 17:22:19.193572 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:19 crc kubenswrapper[4989]: I1213 17:22:19.385282 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:19 crc kubenswrapper[4989]: I1213 17:22:19.385939 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:19 crc kubenswrapper[4989]: I1213 17:22:19.425411 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:19 crc kubenswrapper[4989]: I1213 17:22:19.637422 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6bhwp" Dec 13 17:22:20 crc kubenswrapper[4989]: I1213 17:22:20.238341 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xsmj4" podUID="1e515c25-23f4-49f2-bc3f-0fbb5e8046be" containerName="registry-server" probeResult="failure" output=< Dec 13 17:22:20 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Dec 13 17:22:20 crc kubenswrapper[4989]: > Dec 13 17:22:29 crc kubenswrapper[4989]: I1213 17:22:29.251209 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:29 crc kubenswrapper[4989]: I1213 17:22:29.322759 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xsmj4" Dec 13 17:22:38 crc kubenswrapper[4989]: I1213 17:22:38.425990 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" podUID="fca2d650-00f0-4f7e-9735-540fe4d2a0e0" containerName="registry" containerID="cri-o://1c9fe4feadc2a115ba69b9808cc226433034f03a80e262e53cb0e1ff8311fd28" gracePeriod=30 Dec 13 17:22:38 crc kubenswrapper[4989]: I1213 17:22:38.704833 4989 generic.go:334] "Generic (PLEG): container finished" podID="fca2d650-00f0-4f7e-9735-540fe4d2a0e0" containerID="1c9fe4feadc2a115ba69b9808cc226433034f03a80e262e53cb0e1ff8311fd28" exitCode=0 Dec 13 17:22:38 crc kubenswrapper[4989]: I1213 17:22:38.705123 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" event={"ID":"fca2d650-00f0-4f7e-9735-540fe4d2a0e0","Type":"ContainerDied","Data":"1c9fe4feadc2a115ba69b9808cc226433034f03a80e262e53cb0e1ff8311fd28"} Dec 13 17:22:38 crc kubenswrapper[4989]: I1213 17:22:38.829899 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:22:38 crc kubenswrapper[4989]: I1213 17:22:38.830068 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:22:38 crc kubenswrapper[4989]: I1213 17:22:38.856832 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.033367 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-installation-pull-secrets\") pod \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.033740 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-certificates\") pod \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.034643 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "fca2d650-00f0-4f7e-9735-540fe4d2a0e0" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.033860 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-ca-trust-extracted\") pod \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.034732 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-trusted-ca\") pod \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.034874 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.034908 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-bound-sa-token\") pod \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.034938 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ltx5\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-kube-api-access-6ltx5\") pod \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.034969 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-tls\") pod \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\" (UID: \"fca2d650-00f0-4f7e-9735-540fe4d2a0e0\") " Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.035664 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "fca2d650-00f0-4f7e-9735-540fe4d2a0e0" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.036310 4989 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.036343 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.040926 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "fca2d650-00f0-4f7e-9735-540fe4d2a0e0" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.040963 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "fca2d650-00f0-4f7e-9735-540fe4d2a0e0" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.045632 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-kube-api-access-6ltx5" (OuterVolumeSpecName: "kube-api-access-6ltx5") pod "fca2d650-00f0-4f7e-9735-540fe4d2a0e0" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0"). InnerVolumeSpecName "kube-api-access-6ltx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.050069 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "fca2d650-00f0-4f7e-9735-540fe4d2a0e0" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.050192 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "fca2d650-00f0-4f7e-9735-540fe4d2a0e0" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.050862 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "fca2d650-00f0-4f7e-9735-540fe4d2a0e0" (UID: "fca2d650-00f0-4f7e-9735-540fe4d2a0e0"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.137445 4989 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.137478 4989 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.137487 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ltx5\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-kube-api-access-6ltx5\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.137498 4989 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.137512 4989 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fca2d650-00f0-4f7e-9735-540fe4d2a0e0-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.713336 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" event={"ID":"fca2d650-00f0-4f7e-9735-540fe4d2a0e0","Type":"ContainerDied","Data":"e28856604f0c54f3a335e61e671ec16e53868fa7ff88e4d4d2a376c704241490"} Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.713390 4989 scope.go:117] "RemoveContainer" containerID="1c9fe4feadc2a115ba69b9808cc226433034f03a80e262e53cb0e1ff8311fd28" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.713443 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-9blxx" Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.761547 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9blxx"] Dec 13 17:22:39 crc kubenswrapper[4989]: I1213 17:22:39.766617 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-9blxx"] Dec 13 17:22:40 crc kubenswrapper[4989]: I1213 17:22:40.022223 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fca2d650-00f0-4f7e-9735-540fe4d2a0e0" path="/var/lib/kubelet/pods/fca2d650-00f0-4f7e-9735-540fe4d2a0e0/volumes" Dec 13 17:23:08 crc kubenswrapper[4989]: I1213 17:23:08.829456 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:23:08 crc kubenswrapper[4989]: I1213 17:23:08.830094 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:23:38 crc kubenswrapper[4989]: I1213 17:23:38.829698 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:23:38 crc kubenswrapper[4989]: I1213 17:23:38.830266 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:23:38 crc kubenswrapper[4989]: I1213 17:23:38.830321 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:23:38 crc kubenswrapper[4989]: I1213 17:23:38.830881 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1707457e30b3e0348e549e5cd99a4006766e7e37c627bd9d4764bc772d5280c9"} pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 17:23:38 crc kubenswrapper[4989]: I1213 17:23:38.830935 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" containerID="cri-o://1707457e30b3e0348e549e5cd99a4006766e7e37c627bd9d4764bc772d5280c9" gracePeriod=600 Dec 13 17:23:39 crc kubenswrapper[4989]: I1213 17:23:39.068283 4989 generic.go:334] "Generic (PLEG): container finished" podID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerID="1707457e30b3e0348e549e5cd99a4006766e7e37c627bd9d4764bc772d5280c9" exitCode=0 Dec 13 17:23:39 crc kubenswrapper[4989]: I1213 17:23:39.068353 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerDied","Data":"1707457e30b3e0348e549e5cd99a4006766e7e37c627bd9d4764bc772d5280c9"} Dec 13 17:23:39 crc kubenswrapper[4989]: I1213 17:23:39.068556 4989 scope.go:117] "RemoveContainer" containerID="8f5e58e94fad4abe3b881e8095584196f669443fca2e1e83788f2c135da70ce0" Dec 13 17:23:40 crc kubenswrapper[4989]: I1213 17:23:40.078661 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerStarted","Data":"c77b5f0a81370cabedf086384605132d2ee7a5c7c37053b21e73aede752d2782"} Dec 13 17:25:26 crc kubenswrapper[4989]: I1213 17:25:26.224599 4989 scope.go:117] "RemoveContainer" containerID="95d192dd0234f05a0d70de672c617494fcc159e617838b5d8cbd6ed9f463d1c5" Dec 13 17:26:08 crc kubenswrapper[4989]: I1213 17:26:08.829362 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:26:08 crc kubenswrapper[4989]: I1213 17:26:08.830184 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:26:26 crc kubenswrapper[4989]: I1213 17:26:26.255282 4989 scope.go:117] "RemoveContainer" containerID="5f1539dd6b5d87185cf413a4417022f4e2a9582cc32f9c65896c21d01ac271ae" Dec 13 17:26:26 crc kubenswrapper[4989]: I1213 17:26:26.274567 4989 scope.go:117] "RemoveContainer" containerID="e64b649cc832ab275c1d65a709fa635c790e86e702c1a466cb5608811cd28215" Dec 13 17:26:38 crc kubenswrapper[4989]: I1213 17:26:38.830027 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:26:38 crc kubenswrapper[4989]: I1213 17:26:38.831125 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:27:08 crc kubenswrapper[4989]: I1213 17:27:08.829145 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:27:08 crc kubenswrapper[4989]: I1213 17:27:08.829680 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:27:08 crc kubenswrapper[4989]: I1213 17:27:08.829728 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:27:08 crc kubenswrapper[4989]: I1213 17:27:08.830363 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c77b5f0a81370cabedf086384605132d2ee7a5c7c37053b21e73aede752d2782"} pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 17:27:08 crc kubenswrapper[4989]: I1213 17:27:08.830413 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" containerID="cri-o://c77b5f0a81370cabedf086384605132d2ee7a5c7c37053b21e73aede752d2782" gracePeriod=600 Dec 13 17:27:09 crc kubenswrapper[4989]: I1213 17:27:09.316763 4989 generic.go:334] "Generic (PLEG): container finished" podID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerID="c77b5f0a81370cabedf086384605132d2ee7a5c7c37053b21e73aede752d2782" exitCode=0 Dec 13 17:27:09 crc kubenswrapper[4989]: I1213 17:27:09.317039 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerDied","Data":"c77b5f0a81370cabedf086384605132d2ee7a5c7c37053b21e73aede752d2782"} Dec 13 17:27:09 crc kubenswrapper[4989]: I1213 17:27:09.317203 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerStarted","Data":"b6113f2a3ccfb4b50af5284bdcbd31c249a42a87c9815c957aa5451a83308bc9"} Dec 13 17:27:09 crc kubenswrapper[4989]: I1213 17:27:09.317268 4989 scope.go:117] "RemoveContainer" containerID="1707457e30b3e0348e549e5cd99a4006766e7e37c627bd9d4764bc772d5280c9" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.135197 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-z7249"] Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.136300 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovn-controller" containerID="cri-o://b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c" gracePeriod=30 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.136706 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="sbdb" containerID="cri-o://8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01" gracePeriod=30 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.136812 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="nbdb" containerID="cri-o://66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8" gracePeriod=30 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.136862 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="northd" containerID="cri-o://25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9" gracePeriod=30 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.136910 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795" gracePeriod=30 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.136957 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="kube-rbac-proxy-node" containerID="cri-o://48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187" gracePeriod=30 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.137001 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovn-acl-logging" containerID="cri-o://cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859" gracePeriod=30 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.176499 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" containerID="cri-o://14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa" gracePeriod=30 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.422555 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovnkube-controller/3.log" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.425398 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovn-acl-logging/0.log" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426121 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovn-controller/0.log" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426504 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa" exitCode=0 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426530 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01" exitCode=0 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426537 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8" exitCode=0 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426543 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795" exitCode=0 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426552 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187" exitCode=0 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426560 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859" exitCode=143 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426567 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c" exitCode=143 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426582 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa"} Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426620 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01"} Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426636 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8"} Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426648 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795"} Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426658 4989 scope.go:117] "RemoveContainer" containerID="8921dfdb7cfb850fe16d9cce4c04fa1c0bcd7132a20d34bfa957b68c79d0d4c1" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426665 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187"} Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426746 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859"} Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.426763 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c"} Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.428302 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hllvq_1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf/kube-multus/2.log" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.428870 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hllvq_1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf/kube-multus/1.log" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.428907 4989 generic.go:334] "Generic (PLEG): container finished" podID="1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf" containerID="115cb6d4fb5c16108922c082c15d8ae751144a27bf4c118f70fbfceb4d6c3964" exitCode=2 Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.428944 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hllvq" event={"ID":"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf","Type":"ContainerDied","Data":"115cb6d4fb5c16108922c082c15d8ae751144a27bf4c118f70fbfceb4d6c3964"} Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.429456 4989 scope.go:117] "RemoveContainer" containerID="115cb6d4fb5c16108922c082c15d8ae751144a27bf4c118f70fbfceb4d6c3964" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.429776 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-hllvq_openshift-multus(1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf)\"" pod="openshift-multus/multus-hllvq" podUID="1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.477323 4989 scope.go:117] "RemoveContainer" containerID="4a4cdcbda385c1dd32d5e22ce8a6ef14098ed09f2422691e3c00cc817315bf5c" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.491996 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovn-acl-logging/0.log" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.492771 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovn-controller/0.log" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.493297 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531017 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-slash\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531069 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-script-lib\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531088 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-etc-openvswitch\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531104 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-bin\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531122 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-config\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531144 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-var-lib-openvswitch\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531159 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-log-socket\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531174 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-systemd-units\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531190 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-netns\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531210 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blhj7\" (UniqueName: \"kubernetes.io/projected/101724b9-153f-4f9d-849a-c04a343e7446-kube-api-access-blhj7\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531228 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-node-log\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531245 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-systemd\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531267 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-kubelet\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531692 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531984 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-log-socket" (OuterVolumeSpecName: "log-socket") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531783 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531821 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531835 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-slash" (OuterVolumeSpecName: "host-slash") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532046 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532086 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532095 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532264 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532297 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532519 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-node-log" (OuterVolumeSpecName: "node-log") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.531733 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-ovn-kubernetes\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532578 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-env-overrides\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532597 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-ovn\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532613 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-var-lib-cni-networks-ovn-kubernetes\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532634 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-netd\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532652 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/101724b9-153f-4f9d-849a-c04a343e7446-ovn-node-metrics-cert\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532671 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-openvswitch\") pod \"101724b9-153f-4f9d-849a-c04a343e7446\" (UID: \"101724b9-153f-4f9d-849a-c04a343e7446\") " Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532694 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532700 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532729 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532754 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532922 4989 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532934 4989 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-slash\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532943 4989 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532952 4989 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532961 4989 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532973 4989 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532984 4989 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.532992 4989 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-log-socket\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.533001 4989 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.533011 4989 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.533019 4989 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-node-log\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.533028 4989 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.533036 4989 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.533044 4989 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.533055 4989 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.533133 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.533530 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.543147 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/101724b9-153f-4f9d-849a-c04a343e7446-kube-api-access-blhj7" (OuterVolumeSpecName: "kube-api-access-blhj7") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "kube-api-access-blhj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.543458 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/101724b9-153f-4f9d-849a-c04a343e7446-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.551688 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f6xbk"] Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.551966 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="kubecfg-setup" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.551993 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="kubecfg-setup" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552005 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="northd" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552014 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="northd" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552026 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="sbdb" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552035 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="sbdb" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552049 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="nbdb" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552059 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="nbdb" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552072 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552080 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552090 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552098 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552107 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552115 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552129 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="kube-rbac-proxy-ovn-metrics" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552139 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="kube-rbac-proxy-ovn-metrics" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552152 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552161 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552174 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovn-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552184 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovn-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552200 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovn-acl-logging" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552212 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovn-acl-logging" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552223 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="kube-rbac-proxy-node" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552233 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="kube-rbac-proxy-node" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552249 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fca2d650-00f0-4f7e-9735-540fe4d2a0e0" containerName="registry" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552260 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fca2d650-00f0-4f7e-9735-540fe4d2a0e0" containerName="registry" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552371 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="northd" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552386 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="sbdb" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552396 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fca2d650-00f0-4f7e-9735-540fe4d2a0e0" containerName="registry" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552406 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovn-acl-logging" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552418 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552427 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="nbdb" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552438 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552449 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="kube-rbac-proxy-ovn-metrics" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552462 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552473 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovn-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552484 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="kube-rbac-proxy-node" Dec 13 17:27:27 crc kubenswrapper[4989]: E1213 17:27:27.552599 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552609 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552726 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.552739 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="101724b9-153f-4f9d-849a-c04a343e7446" containerName="ovnkube-controller" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.554864 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "101724b9-153f-4f9d-849a-c04a343e7446" (UID: "101724b9-153f-4f9d-849a-c04a343e7446"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.555187 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.634672 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-systemd-units\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.634731 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-etc-openvswitch\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.634762 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-kubelet\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.634782 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-run-systemd\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.634824 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-log-socket\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.634846 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-node-log\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.634873 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.634899 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-run-ovn\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635144 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-var-lib-openvswitch\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635220 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-slash\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635259 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/673f7904-ed03-47ef-95ee-05f0e8ab50ca-ovnkube-script-lib\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635298 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-cni-netd\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635340 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-run-openvswitch\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635380 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/673f7904-ed03-47ef-95ee-05f0e8ab50ca-env-overrides\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635423 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdtb9\" (UniqueName: \"kubernetes.io/projected/673f7904-ed03-47ef-95ee-05f0e8ab50ca-kube-api-access-xdtb9\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635541 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-run-netns\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635769 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/673f7904-ed03-47ef-95ee-05f0e8ab50ca-ovn-node-metrics-cert\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635853 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-run-ovn-kubernetes\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635916 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/673f7904-ed03-47ef-95ee-05f0e8ab50ca-ovnkube-config\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.635947 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-cni-bin\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.636020 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blhj7\" (UniqueName: \"kubernetes.io/projected/101724b9-153f-4f9d-849a-c04a343e7446-kube-api-access-blhj7\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.636038 4989 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.636051 4989 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/101724b9-153f-4f9d-849a-c04a343e7446-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.636061 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/101724b9-153f-4f9d-849a-c04a343e7446-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.636069 4989 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/101724b9-153f-4f9d-849a-c04a343e7446-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737220 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737257 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-run-ovn\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737279 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-var-lib-openvswitch\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737299 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-slash\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737315 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/673f7904-ed03-47ef-95ee-05f0e8ab50ca-ovnkube-script-lib\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737332 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-cni-netd\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737339 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737363 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-var-lib-openvswitch\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737380 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-run-ovn\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737423 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-run-openvswitch\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737445 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/673f7904-ed03-47ef-95ee-05f0e8ab50ca-env-overrides\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737480 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-cni-netd\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737487 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdtb9\" (UniqueName: \"kubernetes.io/projected/673f7904-ed03-47ef-95ee-05f0e8ab50ca-kube-api-access-xdtb9\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737487 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-run-openvswitch\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737535 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-run-netns\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737570 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/673f7904-ed03-47ef-95ee-05f0e8ab50ca-ovn-node-metrics-cert\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737592 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-run-ovn-kubernetes\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737618 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/673f7904-ed03-47ef-95ee-05f0e8ab50ca-ovnkube-config\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737640 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-cni-bin\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737663 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-systemd-units\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737689 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-etc-openvswitch\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737716 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-kubelet\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737738 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-log-socket\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737760 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-run-systemd\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737783 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-node-log\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737931 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-node-log\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.737967 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-run-ovn-kubernetes\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.738087 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-log-socket\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.738122 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-kubelet\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.738131 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/673f7904-ed03-47ef-95ee-05f0e8ab50ca-env-overrides\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.738148 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-cni-bin\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.738170 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-run-netns\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.738150 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-etc-openvswitch\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.738231 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-systemd-units\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.738264 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-run-systemd\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.738320 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/673f7904-ed03-47ef-95ee-05f0e8ab50ca-host-slash\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.738287 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/673f7904-ed03-47ef-95ee-05f0e8ab50ca-ovnkube-script-lib\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.738459 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/673f7904-ed03-47ef-95ee-05f0e8ab50ca-ovnkube-config\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.740992 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/673f7904-ed03-47ef-95ee-05f0e8ab50ca-ovn-node-metrics-cert\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.754663 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdtb9\" (UniqueName: \"kubernetes.io/projected/673f7904-ed03-47ef-95ee-05f0e8ab50ca-kube-api-access-xdtb9\") pod \"ovnkube-node-f6xbk\" (UID: \"673f7904-ed03-47ef-95ee-05f0e8ab50ca\") " pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: I1213 17:27:27.875765 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:27 crc kubenswrapper[4989]: W1213 17:27:27.905434 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod673f7904_ed03_47ef_95ee_05f0e8ab50ca.slice/crio-fff5d1f02b63a5e6d6e077b1509a7df2731bb124355fa0fe056234e7a420a86a WatchSource:0}: Error finding container fff5d1f02b63a5e6d6e077b1509a7df2731bb124355fa0fe056234e7a420a86a: Status 404 returned error can't find the container with id fff5d1f02b63a5e6d6e077b1509a7df2731bb124355fa0fe056234e7a420a86a Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.437298 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hllvq_1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf/kube-multus/2.log" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.438994 4989 generic.go:334] "Generic (PLEG): container finished" podID="673f7904-ed03-47ef-95ee-05f0e8ab50ca" containerID="9d60ef5b8973141bdb9725143707170e27ce2911f7ddb4e3e22cb028ab4f3aa4" exitCode=0 Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.439088 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" event={"ID":"673f7904-ed03-47ef-95ee-05f0e8ab50ca","Type":"ContainerDied","Data":"9d60ef5b8973141bdb9725143707170e27ce2911f7ddb4e3e22cb028ab4f3aa4"} Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.439117 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" event={"ID":"673f7904-ed03-47ef-95ee-05f0e8ab50ca","Type":"ContainerStarted","Data":"fff5d1f02b63a5e6d6e077b1509a7df2731bb124355fa0fe056234e7a420a86a"} Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.445116 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovn-acl-logging/0.log" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.447129 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-z7249_101724b9-153f-4f9d-849a-c04a343e7446/ovn-controller/0.log" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.447760 4989 generic.go:334] "Generic (PLEG): container finished" podID="101724b9-153f-4f9d-849a-c04a343e7446" containerID="25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9" exitCode=0 Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.447824 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9"} Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.447857 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" event={"ID":"101724b9-153f-4f9d-849a-c04a343e7446","Type":"ContainerDied","Data":"bafb9bdcc89878b80e073d2652fbd3ea185f04e42ad8f704985f0187ceea4b27"} Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.447876 4989 scope.go:117] "RemoveContainer" containerID="14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.447936 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z7249" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.473562 4989 scope.go:117] "RemoveContainer" containerID="8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.496032 4989 scope.go:117] "RemoveContainer" containerID="66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.527373 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-z7249"] Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.530339 4989 scope.go:117] "RemoveContainer" containerID="25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.565577 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-z7249"] Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.566005 4989 scope.go:117] "RemoveContainer" containerID="d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.580751 4989 scope.go:117] "RemoveContainer" containerID="48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.603802 4989 scope.go:117] "RemoveContainer" containerID="cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.619864 4989 scope.go:117] "RemoveContainer" containerID="b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.649274 4989 scope.go:117] "RemoveContainer" containerID="d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.673445 4989 scope.go:117] "RemoveContainer" containerID="14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa" Dec 13 17:27:28 crc kubenswrapper[4989]: E1213 17:27:28.676526 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa\": container with ID starting with 14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa not found: ID does not exist" containerID="14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.676574 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa"} err="failed to get container status \"14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa\": rpc error: code = NotFound desc = could not find container \"14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa\": container with ID starting with 14b4b9170cc79fa1f57340aacabfc8014f1a44a660d3a459276721caf77007aa not found: ID does not exist" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.676600 4989 scope.go:117] "RemoveContainer" containerID="8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01" Dec 13 17:27:28 crc kubenswrapper[4989]: E1213 17:27:28.676977 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\": container with ID starting with 8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01 not found: ID does not exist" containerID="8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.676995 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01"} err="failed to get container status \"8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\": rpc error: code = NotFound desc = could not find container \"8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01\": container with ID starting with 8050ab7402e18be4ab3a6fdbe1d86d2ed2fcf6ee8981c0d2b343e07f27661c01 not found: ID does not exist" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.677007 4989 scope.go:117] "RemoveContainer" containerID="66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8" Dec 13 17:27:28 crc kubenswrapper[4989]: E1213 17:27:28.677176 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\": container with ID starting with 66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8 not found: ID does not exist" containerID="66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.677195 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8"} err="failed to get container status \"66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\": rpc error: code = NotFound desc = could not find container \"66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8\": container with ID starting with 66771808a88aea610da5ce3b33c9c4a7c1a8e9cba5d069dcea7c568ca3e1c0c8 not found: ID does not exist" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.677206 4989 scope.go:117] "RemoveContainer" containerID="25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9" Dec 13 17:27:28 crc kubenswrapper[4989]: E1213 17:27:28.677383 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\": container with ID starting with 25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9 not found: ID does not exist" containerID="25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.677408 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9"} err="failed to get container status \"25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\": rpc error: code = NotFound desc = could not find container \"25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9\": container with ID starting with 25de10dc375ef809d22603d06e70c981ddb006613533908f7122c4a42f8ad5b9 not found: ID does not exist" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.677422 4989 scope.go:117] "RemoveContainer" containerID="d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795" Dec 13 17:27:28 crc kubenswrapper[4989]: E1213 17:27:28.677596 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\": container with ID starting with d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795 not found: ID does not exist" containerID="d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.677616 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795"} err="failed to get container status \"d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\": rpc error: code = NotFound desc = could not find container \"d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795\": container with ID starting with d0df7e6f19bab9dccb1146f44a4c5278e39b76265aa43dfe3a81a002354f8795 not found: ID does not exist" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.677639 4989 scope.go:117] "RemoveContainer" containerID="48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187" Dec 13 17:27:28 crc kubenswrapper[4989]: E1213 17:27:28.677869 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\": container with ID starting with 48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187 not found: ID does not exist" containerID="48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.677887 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187"} err="failed to get container status \"48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\": rpc error: code = NotFound desc = could not find container \"48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187\": container with ID starting with 48258e5e3c0ec54fd835a21b9fd7462c14d2eafde7a5dfe450d6879eee26d187 not found: ID does not exist" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.677899 4989 scope.go:117] "RemoveContainer" containerID="cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859" Dec 13 17:27:28 crc kubenswrapper[4989]: E1213 17:27:28.678052 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\": container with ID starting with cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859 not found: ID does not exist" containerID="cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.678070 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859"} err="failed to get container status \"cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\": rpc error: code = NotFound desc = could not find container \"cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859\": container with ID starting with cd96e7f4b212bf14af11961672c96d7b0948b1566f9fd6614c31a114efeba859 not found: ID does not exist" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.678081 4989 scope.go:117] "RemoveContainer" containerID="b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c" Dec 13 17:27:28 crc kubenswrapper[4989]: E1213 17:27:28.678221 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\": container with ID starting with b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c not found: ID does not exist" containerID="b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.678236 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c"} err="failed to get container status \"b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\": rpc error: code = NotFound desc = could not find container \"b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c\": container with ID starting with b85f168441cd32b943e35b869aaf1ae1c32632a6cac34e7e504c3b02f5aaac1c not found: ID does not exist" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.678247 4989 scope.go:117] "RemoveContainer" containerID="d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671" Dec 13 17:27:28 crc kubenswrapper[4989]: E1213 17:27:28.678389 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\": container with ID starting with d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671 not found: ID does not exist" containerID="d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671" Dec 13 17:27:28 crc kubenswrapper[4989]: I1213 17:27:28.678405 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671"} err="failed to get container status \"d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\": rpc error: code = NotFound desc = could not find container \"d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671\": container with ID starting with d21377d683a9cdcc3175549f95f939515f15d2ffc8a1b4433968726b4e32c671 not found: ID does not exist" Dec 13 17:27:29 crc kubenswrapper[4989]: I1213 17:27:29.455375 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" event={"ID":"673f7904-ed03-47ef-95ee-05f0e8ab50ca","Type":"ContainerStarted","Data":"d8eecd7f2479d5fac8dfd0c904627220a092a48d783ecf0be3d6ecbf15539939"} Dec 13 17:27:29 crc kubenswrapper[4989]: I1213 17:27:29.455719 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" event={"ID":"673f7904-ed03-47ef-95ee-05f0e8ab50ca","Type":"ContainerStarted","Data":"8fa1058b040a9947587373c8d232149d98bdafc97981b6a3ac9fb8ca36ce50c7"} Dec 13 17:27:29 crc kubenswrapper[4989]: I1213 17:27:29.455733 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" event={"ID":"673f7904-ed03-47ef-95ee-05f0e8ab50ca","Type":"ContainerStarted","Data":"7f3fee18597c9cef665d5e2bcf9be0281987df5dbccf8d4c3a92dc0615eee939"} Dec 13 17:27:29 crc kubenswrapper[4989]: I1213 17:27:29.455744 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" event={"ID":"673f7904-ed03-47ef-95ee-05f0e8ab50ca","Type":"ContainerStarted","Data":"a3c5f93baebfd46e65a96b456bdfcd7c0428a0ab8e5597c3ff5614c6045f2e4b"} Dec 13 17:27:29 crc kubenswrapper[4989]: I1213 17:27:29.455755 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" event={"ID":"673f7904-ed03-47ef-95ee-05f0e8ab50ca","Type":"ContainerStarted","Data":"4415de56d824656c5f1833e25e79828ab5f363e6eab132aac9926ff6ab719910"} Dec 13 17:27:29 crc kubenswrapper[4989]: I1213 17:27:29.455766 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" event={"ID":"673f7904-ed03-47ef-95ee-05f0e8ab50ca","Type":"ContainerStarted","Data":"bda90fe031ce566c827d7acebeaa86e86a1a7f6fc558f357444da4e04a23a296"} Dec 13 17:27:30 crc kubenswrapper[4989]: I1213 17:27:30.026706 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="101724b9-153f-4f9d-849a-c04a343e7446" path="/var/lib/kubelet/pods/101724b9-153f-4f9d-849a-c04a343e7446/volumes" Dec 13 17:27:31 crc kubenswrapper[4989]: I1213 17:27:31.476244 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" event={"ID":"673f7904-ed03-47ef-95ee-05f0e8ab50ca","Type":"ContainerStarted","Data":"a152a143f7588f71a36a3b7b043d053010d0f5f5ef47ae62fb9d511b80ca4128"} Dec 13 17:27:34 crc kubenswrapper[4989]: I1213 17:27:34.500349 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" event={"ID":"673f7904-ed03-47ef-95ee-05f0e8ab50ca","Type":"ContainerStarted","Data":"9c71cddf17b8fdcbc24daa4c269e7cd77486ce08d5881b152be31da832d3e2e0"} Dec 13 17:27:34 crc kubenswrapper[4989]: I1213 17:27:34.500874 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:34 crc kubenswrapper[4989]: I1213 17:27:34.500904 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:34 crc kubenswrapper[4989]: I1213 17:27:34.500917 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:34 crc kubenswrapper[4989]: I1213 17:27:34.527538 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:34 crc kubenswrapper[4989]: I1213 17:27:34.531149 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:34 crc kubenswrapper[4989]: I1213 17:27:34.542427 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" podStartSLOduration=7.542414115 podStartE2EDuration="7.542414115s" podCreationTimestamp="2025-12-13 17:27:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:27:34.540485134 +0000 UTC m=+669.146932272" watchObservedRunningTime="2025-12-13 17:27:34.542414115 +0000 UTC m=+669.148861253" Dec 13 17:27:38 crc kubenswrapper[4989]: I1213 17:27:38.015418 4989 scope.go:117] "RemoveContainer" containerID="115cb6d4fb5c16108922c082c15d8ae751144a27bf4c118f70fbfceb4d6c3964" Dec 13 17:27:38 crc kubenswrapper[4989]: E1213 17:27:38.016211 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-hllvq_openshift-multus(1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf)\"" pod="openshift-multus/multus-hllvq" podUID="1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf" Dec 13 17:27:49 crc kubenswrapper[4989]: I1213 17:27:49.014363 4989 scope.go:117] "RemoveContainer" containerID="115cb6d4fb5c16108922c082c15d8ae751144a27bf4c118f70fbfceb4d6c3964" Dec 13 17:27:49 crc kubenswrapper[4989]: I1213 17:27:49.580490 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hllvq_1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf/kube-multus/2.log" Dec 13 17:27:49 crc kubenswrapper[4989]: I1213 17:27:49.580770 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hllvq" event={"ID":"1ea718eb-ab21-4f3c-8d0d-c6cf4ffe69bf","Type":"ContainerStarted","Data":"a5fcc8d4788282392e50e17f29cd476fb6129b3ce2d2a22180964c16e706441d"} Dec 13 17:27:51 crc kubenswrapper[4989]: I1213 17:27:51.940954 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc"] Dec 13 17:27:51 crc kubenswrapper[4989]: I1213 17:27:51.942232 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:51 crc kubenswrapper[4989]: I1213 17:27:51.946616 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc"] Dec 13 17:27:51 crc kubenswrapper[4989]: I1213 17:27:51.947451 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.078274 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t94m4\" (UniqueName: \"kubernetes.io/projected/d2d02100-c895-4311-955b-a06f5d7b1e0d-kube-api-access-t94m4\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.078360 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.078592 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.180493 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t94m4\" (UniqueName: \"kubernetes.io/projected/d2d02100-c895-4311-955b-a06f5d7b1e0d-kube-api-access-t94m4\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.180547 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.180601 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.181054 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.181080 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.197423 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t94m4\" (UniqueName: \"kubernetes.io/projected/d2d02100-c895-4311-955b-a06f5d7b1e0d-kube-api-access-t94m4\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.257293 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.440681 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc"] Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.596411 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" event={"ID":"d2d02100-c895-4311-955b-a06f5d7b1e0d","Type":"ContainerStarted","Data":"e5b519267a9d14d6bc97db4e9387091551927d050c2f28ce8e867b97cd63b41a"} Dec 13 17:27:52 crc kubenswrapper[4989]: I1213 17:27:52.596455 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" event={"ID":"d2d02100-c895-4311-955b-a06f5d7b1e0d","Type":"ContainerStarted","Data":"c11e22e50b323e09eebdc3a7178921157ce137476bafac227266342ecc415dcf"} Dec 13 17:27:53 crc kubenswrapper[4989]: I1213 17:27:53.604872 4989 generic.go:334] "Generic (PLEG): container finished" podID="d2d02100-c895-4311-955b-a06f5d7b1e0d" containerID="e5b519267a9d14d6bc97db4e9387091551927d050c2f28ce8e867b97cd63b41a" exitCode=0 Dec 13 17:27:53 crc kubenswrapper[4989]: I1213 17:27:53.604953 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" event={"ID":"d2d02100-c895-4311-955b-a06f5d7b1e0d","Type":"ContainerDied","Data":"e5b519267a9d14d6bc97db4e9387091551927d050c2f28ce8e867b97cd63b41a"} Dec 13 17:27:53 crc kubenswrapper[4989]: I1213 17:27:53.606535 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 17:27:55 crc kubenswrapper[4989]: I1213 17:27:55.618332 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" event={"ID":"d2d02100-c895-4311-955b-a06f5d7b1e0d","Type":"ContainerStarted","Data":"6700d53bc33581df486c3fcc56bc496c7588e80a1096520bf2db68ab1f2e672c"} Dec 13 17:27:56 crc kubenswrapper[4989]: I1213 17:27:56.627394 4989 generic.go:334] "Generic (PLEG): container finished" podID="d2d02100-c895-4311-955b-a06f5d7b1e0d" containerID="6700d53bc33581df486c3fcc56bc496c7588e80a1096520bf2db68ab1f2e672c" exitCode=0 Dec 13 17:27:56 crc kubenswrapper[4989]: I1213 17:27:56.627764 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" event={"ID":"d2d02100-c895-4311-955b-a06f5d7b1e0d","Type":"ContainerDied","Data":"6700d53bc33581df486c3fcc56bc496c7588e80a1096520bf2db68ab1f2e672c"} Dec 13 17:27:57 crc kubenswrapper[4989]: I1213 17:27:57.638112 4989 generic.go:334] "Generic (PLEG): container finished" podID="d2d02100-c895-4311-955b-a06f5d7b1e0d" containerID="a1ea1de7d1f899c044e3086781cd1cf92ff40777d4ca5744103ae85745c17635" exitCode=0 Dec 13 17:27:57 crc kubenswrapper[4989]: I1213 17:27:57.638280 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" event={"ID":"d2d02100-c895-4311-955b-a06f5d7b1e0d","Type":"ContainerDied","Data":"a1ea1de7d1f899c044e3086781cd1cf92ff40777d4ca5744103ae85745c17635"} Dec 13 17:27:57 crc kubenswrapper[4989]: I1213 17:27:57.895720 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f6xbk" Dec 13 17:27:58 crc kubenswrapper[4989]: I1213 17:27:58.896081 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:27:58 crc kubenswrapper[4989]: I1213 17:27:58.967503 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-util\") pod \"d2d02100-c895-4311-955b-a06f5d7b1e0d\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " Dec 13 17:27:58 crc kubenswrapper[4989]: I1213 17:27:58.967589 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t94m4\" (UniqueName: \"kubernetes.io/projected/d2d02100-c895-4311-955b-a06f5d7b1e0d-kube-api-access-t94m4\") pod \"d2d02100-c895-4311-955b-a06f5d7b1e0d\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " Dec 13 17:27:58 crc kubenswrapper[4989]: I1213 17:27:58.967618 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-bundle\") pod \"d2d02100-c895-4311-955b-a06f5d7b1e0d\" (UID: \"d2d02100-c895-4311-955b-a06f5d7b1e0d\") " Dec 13 17:27:58 crc kubenswrapper[4989]: I1213 17:27:58.968585 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-bundle" (OuterVolumeSpecName: "bundle") pod "d2d02100-c895-4311-955b-a06f5d7b1e0d" (UID: "d2d02100-c895-4311-955b-a06f5d7b1e0d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:27:58 crc kubenswrapper[4989]: I1213 17:27:58.975977 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d02100-c895-4311-955b-a06f5d7b1e0d-kube-api-access-t94m4" (OuterVolumeSpecName: "kube-api-access-t94m4") pod "d2d02100-c895-4311-955b-a06f5d7b1e0d" (UID: "d2d02100-c895-4311-955b-a06f5d7b1e0d"). InnerVolumeSpecName "kube-api-access-t94m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:27:58 crc kubenswrapper[4989]: I1213 17:27:58.977412 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-util" (OuterVolumeSpecName: "util") pod "d2d02100-c895-4311-955b-a06f5d7b1e0d" (UID: "d2d02100-c895-4311-955b-a06f5d7b1e0d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:27:59 crc kubenswrapper[4989]: I1213 17:27:59.070024 4989 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-util\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:59 crc kubenswrapper[4989]: I1213 17:27:59.070337 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t94m4\" (UniqueName: \"kubernetes.io/projected/d2d02100-c895-4311-955b-a06f5d7b1e0d-kube-api-access-t94m4\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:59 crc kubenswrapper[4989]: I1213 17:27:59.070414 4989 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2d02100-c895-4311-955b-a06f5d7b1e0d-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:27:59 crc kubenswrapper[4989]: I1213 17:27:59.651041 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" event={"ID":"d2d02100-c895-4311-955b-a06f5d7b1e0d","Type":"ContainerDied","Data":"c11e22e50b323e09eebdc3a7178921157ce137476bafac227266342ecc415dcf"} Dec 13 17:27:59 crc kubenswrapper[4989]: I1213 17:27:59.651369 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c11e22e50b323e09eebdc3a7178921157ce137476bafac227266342ecc415dcf" Dec 13 17:27:59 crc kubenswrapper[4989]: I1213 17:27:59.651106 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.548300 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q"] Dec 13 17:28:07 crc kubenswrapper[4989]: E1213 17:28:07.550082 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d02100-c895-4311-955b-a06f5d7b1e0d" containerName="extract" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.550165 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d02100-c895-4311-955b-a06f5d7b1e0d" containerName="extract" Dec 13 17:28:07 crc kubenswrapper[4989]: E1213 17:28:07.550229 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d02100-c895-4311-955b-a06f5d7b1e0d" containerName="util" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.550294 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d02100-c895-4311-955b-a06f5d7b1e0d" containerName="util" Dec 13 17:28:07 crc kubenswrapper[4989]: E1213 17:28:07.550359 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d02100-c895-4311-955b-a06f5d7b1e0d" containerName="pull" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.550421 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d02100-c895-4311-955b-a06f5d7b1e0d" containerName="pull" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.550578 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d02100-c895-4311-955b-a06f5d7b1e0d" containerName="extract" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.551132 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.556023 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.556025 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.556126 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.556368 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.556423 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-gjnp8" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.569000 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q"] Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.576527 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ba2b41f-9366-455a-ab5c-51a199ac8d52-webhook-cert\") pod \"metallb-operator-controller-manager-64f9694c74-whn2q\" (UID: \"1ba2b41f-9366-455a-ab5c-51a199ac8d52\") " pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.576593 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ba2b41f-9366-455a-ab5c-51a199ac8d52-apiservice-cert\") pod \"metallb-operator-controller-manager-64f9694c74-whn2q\" (UID: \"1ba2b41f-9366-455a-ab5c-51a199ac8d52\") " pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.576636 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kt95\" (UniqueName: \"kubernetes.io/projected/1ba2b41f-9366-455a-ab5c-51a199ac8d52-kube-api-access-4kt95\") pod \"metallb-operator-controller-manager-64f9694c74-whn2q\" (UID: \"1ba2b41f-9366-455a-ab5c-51a199ac8d52\") " pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.677377 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ba2b41f-9366-455a-ab5c-51a199ac8d52-webhook-cert\") pod \"metallb-operator-controller-manager-64f9694c74-whn2q\" (UID: \"1ba2b41f-9366-455a-ab5c-51a199ac8d52\") " pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.677433 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ba2b41f-9366-455a-ab5c-51a199ac8d52-apiservice-cert\") pod \"metallb-operator-controller-manager-64f9694c74-whn2q\" (UID: \"1ba2b41f-9366-455a-ab5c-51a199ac8d52\") " pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.677476 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kt95\" (UniqueName: \"kubernetes.io/projected/1ba2b41f-9366-455a-ab5c-51a199ac8d52-kube-api-access-4kt95\") pod \"metallb-operator-controller-manager-64f9694c74-whn2q\" (UID: \"1ba2b41f-9366-455a-ab5c-51a199ac8d52\") " pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.685647 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ba2b41f-9366-455a-ab5c-51a199ac8d52-apiservice-cert\") pod \"metallb-operator-controller-manager-64f9694c74-whn2q\" (UID: \"1ba2b41f-9366-455a-ab5c-51a199ac8d52\") " pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.685657 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ba2b41f-9366-455a-ab5c-51a199ac8d52-webhook-cert\") pod \"metallb-operator-controller-manager-64f9694c74-whn2q\" (UID: \"1ba2b41f-9366-455a-ab5c-51a199ac8d52\") " pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.694425 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kt95\" (UniqueName: \"kubernetes.io/projected/1ba2b41f-9366-455a-ab5c-51a199ac8d52-kube-api-access-4kt95\") pod \"metallb-operator-controller-manager-64f9694c74-whn2q\" (UID: \"1ba2b41f-9366-455a-ab5c-51a199ac8d52\") " pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.872388 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.964155 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst"] Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.965168 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.973751 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.975873 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.976281 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-8cgl4" Dec 13 17:28:07 crc kubenswrapper[4989]: I1213 17:28:07.993160 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst"] Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.090403 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jds4d\" (UniqueName: \"kubernetes.io/projected/9f540b88-5e5f-4420-8b2c-d49f42087943-kube-api-access-jds4d\") pod \"metallb-operator-webhook-server-84cdcb4857-zvzst\" (UID: \"9f540b88-5e5f-4420-8b2c-d49f42087943\") " pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.090765 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f540b88-5e5f-4420-8b2c-d49f42087943-webhook-cert\") pod \"metallb-operator-webhook-server-84cdcb4857-zvzst\" (UID: \"9f540b88-5e5f-4420-8b2c-d49f42087943\") " pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.090821 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f540b88-5e5f-4420-8b2c-d49f42087943-apiservice-cert\") pod \"metallb-operator-webhook-server-84cdcb4857-zvzst\" (UID: \"9f540b88-5e5f-4420-8b2c-d49f42087943\") " pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.191599 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f540b88-5e5f-4420-8b2c-d49f42087943-webhook-cert\") pod \"metallb-operator-webhook-server-84cdcb4857-zvzst\" (UID: \"9f540b88-5e5f-4420-8b2c-d49f42087943\") " pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.191671 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f540b88-5e5f-4420-8b2c-d49f42087943-apiservice-cert\") pod \"metallb-operator-webhook-server-84cdcb4857-zvzst\" (UID: \"9f540b88-5e5f-4420-8b2c-d49f42087943\") " pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.191718 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jds4d\" (UniqueName: \"kubernetes.io/projected/9f540b88-5e5f-4420-8b2c-d49f42087943-kube-api-access-jds4d\") pod \"metallb-operator-webhook-server-84cdcb4857-zvzst\" (UID: \"9f540b88-5e5f-4420-8b2c-d49f42087943\") " pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.195702 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f540b88-5e5f-4420-8b2c-d49f42087943-apiservice-cert\") pod \"metallb-operator-webhook-server-84cdcb4857-zvzst\" (UID: \"9f540b88-5e5f-4420-8b2c-d49f42087943\") " pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.196332 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f540b88-5e5f-4420-8b2c-d49f42087943-webhook-cert\") pod \"metallb-operator-webhook-server-84cdcb4857-zvzst\" (UID: \"9f540b88-5e5f-4420-8b2c-d49f42087943\") " pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.209447 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jds4d\" (UniqueName: \"kubernetes.io/projected/9f540b88-5e5f-4420-8b2c-d49f42087943-kube-api-access-jds4d\") pod \"metallb-operator-webhook-server-84cdcb4857-zvzst\" (UID: \"9f540b88-5e5f-4420-8b2c-d49f42087943\") " pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.222489 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q"] Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.279469 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.496190 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst"] Dec 13 17:28:08 crc kubenswrapper[4989]: W1213 17:28:08.501534 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f540b88_5e5f_4420_8b2c_d49f42087943.slice/crio-0bb85f95e9ee623af662e0032301ec8233809d1da8322e8c14afefcdef027df2 WatchSource:0}: Error finding container 0bb85f95e9ee623af662e0032301ec8233809d1da8322e8c14afefcdef027df2: Status 404 returned error can't find the container with id 0bb85f95e9ee623af662e0032301ec8233809d1da8322e8c14afefcdef027df2 Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.698243 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" event={"ID":"9f540b88-5e5f-4420-8b2c-d49f42087943","Type":"ContainerStarted","Data":"0bb85f95e9ee623af662e0032301ec8233809d1da8322e8c14afefcdef027df2"} Dec 13 17:28:08 crc kubenswrapper[4989]: I1213 17:28:08.699191 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" event={"ID":"1ba2b41f-9366-455a-ab5c-51a199ac8d52","Type":"ContainerStarted","Data":"977d9b8979e807659208066b9e2707dedb3e8502a8853373177941443f75e860"} Dec 13 17:28:14 crc kubenswrapper[4989]: I1213 17:28:14.753294 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" event={"ID":"9f540b88-5e5f-4420-8b2c-d49f42087943","Type":"ContainerStarted","Data":"89ec7732d431f1338b1966eacf77cbd9a91de8baa61f835825f51f3c3572fb10"} Dec 13 17:28:14 crc kubenswrapper[4989]: I1213 17:28:14.753838 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:14 crc kubenswrapper[4989]: I1213 17:28:14.756641 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" event={"ID":"1ba2b41f-9366-455a-ab5c-51a199ac8d52","Type":"ContainerStarted","Data":"c27bdf17be1558dac38f29a49633032a2b03a4aba5ddc270cd015ad204ba348b"} Dec 13 17:28:14 crc kubenswrapper[4989]: I1213 17:28:14.756957 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:14 crc kubenswrapper[4989]: I1213 17:28:14.783067 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" podStartSLOduration=1.7724641129999998 podStartE2EDuration="7.783052679s" podCreationTimestamp="2025-12-13 17:28:07 +0000 UTC" firstStartedPulling="2025-12-13 17:28:08.504942284 +0000 UTC m=+703.111389432" lastFinishedPulling="2025-12-13 17:28:14.51553086 +0000 UTC m=+709.121977998" observedRunningTime="2025-12-13 17:28:14.782873974 +0000 UTC m=+709.389321112" watchObservedRunningTime="2025-12-13 17:28:14.783052679 +0000 UTC m=+709.389499817" Dec 13 17:28:14 crc kubenswrapper[4989]: I1213 17:28:14.806315 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" podStartSLOduration=1.528754865 podStartE2EDuration="7.806296702s" podCreationTimestamp="2025-12-13 17:28:07 +0000 UTC" firstStartedPulling="2025-12-13 17:28:08.231472428 +0000 UTC m=+702.837919566" lastFinishedPulling="2025-12-13 17:28:14.509014265 +0000 UTC m=+709.115461403" observedRunningTime="2025-12-13 17:28:14.804506586 +0000 UTC m=+709.410953724" watchObservedRunningTime="2025-12-13 17:28:14.806296702 +0000 UTC m=+709.412743840" Dec 13 17:28:28 crc kubenswrapper[4989]: I1213 17:28:28.284641 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-84cdcb4857-zvzst" Dec 13 17:28:47 crc kubenswrapper[4989]: I1213 17:28:47.876574 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-64f9694c74-whn2q" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.648054 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hq8z9"] Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.650376 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.652405 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-bqv5h" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.652883 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.652989 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.664527 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx"] Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.665215 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.671425 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.707050 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx"] Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.740631 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-r5qp2"] Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.741444 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-r5qp2" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.744763 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-47952" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.744979 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.746920 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.746952 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.755873 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5bddd4b946-55pxp"] Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.756730 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.758510 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.776852 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-55pxp"] Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812211 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-memberlist\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812545 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-metrics-certs\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812628 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c4da269-7f3c-4101-8776-3c10b7379d9c-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-2shrx\" (UID: \"7c4da269-7f3c-4101-8776-3c10b7379d9c\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812670 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzxqw\" (UniqueName: \"kubernetes.io/projected/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-kube-api-access-xzxqw\") pod \"controller-5bddd4b946-55pxp\" (UID: \"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29\") " pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812696 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-metrics\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812724 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-frr-conf\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812838 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-cert\") pod \"controller-5bddd4b946-55pxp\" (UID: \"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29\") " pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812876 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwd7c\" (UniqueName: \"kubernetes.io/projected/931ee02f-2887-4831-a0df-94eccdf81672-kube-api-access-pwd7c\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812903 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsgf9\" (UniqueName: \"kubernetes.io/projected/f2005529-2d5f-4dc3-b783-d30e0e1d7962-kube-api-access-wsgf9\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812952 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f2005529-2d5f-4dc3-b783-d30e0e1d7962-frr-startup\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812967 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/931ee02f-2887-4831-a0df-94eccdf81672-metallb-excludel2\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.812985 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2005529-2d5f-4dc3-b783-d30e0e1d7962-metrics-certs\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.813004 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-metrics-certs\") pod \"controller-5bddd4b946-55pxp\" (UID: \"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29\") " pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.813023 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-frr-sockets\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.813047 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-reloader\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.813089 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsqx7\" (UniqueName: \"kubernetes.io/projected/7c4da269-7f3c-4101-8776-3c10b7379d9c-kube-api-access-hsqx7\") pod \"frr-k8s-webhook-server-7784b6fcf-2shrx\" (UID: \"7c4da269-7f3c-4101-8776-3c10b7379d9c\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914635 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c4da269-7f3c-4101-8776-3c10b7379d9c-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-2shrx\" (UID: \"7c4da269-7f3c-4101-8776-3c10b7379d9c\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914689 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzxqw\" (UniqueName: \"kubernetes.io/projected/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-kube-api-access-xzxqw\") pod \"controller-5bddd4b946-55pxp\" (UID: \"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29\") " pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914714 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-metrics\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914732 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-frr-conf\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914763 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-cert\") pod \"controller-5bddd4b946-55pxp\" (UID: \"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29\") " pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914797 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwd7c\" (UniqueName: \"kubernetes.io/projected/931ee02f-2887-4831-a0df-94eccdf81672-kube-api-access-pwd7c\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914816 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsgf9\" (UniqueName: \"kubernetes.io/projected/f2005529-2d5f-4dc3-b783-d30e0e1d7962-kube-api-access-wsgf9\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914834 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f2005529-2d5f-4dc3-b783-d30e0e1d7962-frr-startup\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914856 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/931ee02f-2887-4831-a0df-94eccdf81672-metallb-excludel2\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914875 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2005529-2d5f-4dc3-b783-d30e0e1d7962-metrics-certs\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914896 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-metrics-certs\") pod \"controller-5bddd4b946-55pxp\" (UID: \"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29\") " pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914932 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-frr-sockets\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914950 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-reloader\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.914981 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsqx7\" (UniqueName: \"kubernetes.io/projected/7c4da269-7f3c-4101-8776-3c10b7379d9c-kube-api-access-hsqx7\") pod \"frr-k8s-webhook-server-7784b6fcf-2shrx\" (UID: \"7c4da269-7f3c-4101-8776-3c10b7379d9c\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.915015 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-memberlist\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.915039 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-metrics-certs\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:48 crc kubenswrapper[4989]: E1213 17:28:48.915154 4989 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 13 17:28:48 crc kubenswrapper[4989]: E1213 17:28:48.915197 4989 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 13 17:28:48 crc kubenswrapper[4989]: E1213 17:28:48.915237 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-metrics-certs podName:c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29 nodeName:}" failed. No retries permitted until 2025-12-13 17:28:49.415212632 +0000 UTC m=+744.021659970 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-metrics-certs") pod "controller-5bddd4b946-55pxp" (UID: "c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29") : secret "controller-certs-secret" not found Dec 13 17:28:48 crc kubenswrapper[4989]: E1213 17:28:48.915265 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-metrics-certs podName:931ee02f-2887-4831-a0df-94eccdf81672 nodeName:}" failed. No retries permitted until 2025-12-13 17:28:49.415252803 +0000 UTC m=+744.021699941 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-metrics-certs") pod "speaker-r5qp2" (UID: "931ee02f-2887-4831-a0df-94eccdf81672") : secret "speaker-certs-secret" not found Dec 13 17:28:48 crc kubenswrapper[4989]: E1213 17:28:48.915320 4989 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 13 17:28:48 crc kubenswrapper[4989]: E1213 17:28:48.915351 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f2005529-2d5f-4dc3-b783-d30e0e1d7962-metrics-certs podName:f2005529-2d5f-4dc3-b783-d30e0e1d7962 nodeName:}" failed. No retries permitted until 2025-12-13 17:28:49.415341696 +0000 UTC m=+744.021789114 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f2005529-2d5f-4dc3-b783-d30e0e1d7962-metrics-certs") pod "frr-k8s-hq8z9" (UID: "f2005529-2d5f-4dc3-b783-d30e0e1d7962") : secret "frr-k8s-certs-secret" not found Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.915372 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-metrics\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.915672 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-frr-conf\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: E1213 17:28:48.915733 4989 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.915747 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-frr-sockets\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: E1213 17:28:48.915776 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-memberlist podName:931ee02f-2887-4831-a0df-94eccdf81672 nodeName:}" failed. No retries permitted until 2025-12-13 17:28:49.415761758 +0000 UTC m=+744.022208896 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-memberlist") pod "speaker-r5qp2" (UID: "931ee02f-2887-4831-a0df-94eccdf81672") : secret "metallb-memberlist" not found Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.915925 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f2005529-2d5f-4dc3-b783-d30e0e1d7962-reloader\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.916186 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/931ee02f-2887-4831-a0df-94eccdf81672-metallb-excludel2\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.916942 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f2005529-2d5f-4dc3-b783-d30e0e1d7962-frr-startup\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.921712 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.935591 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-cert\") pod \"controller-5bddd4b946-55pxp\" (UID: \"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29\") " pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.937754 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7c4da269-7f3c-4101-8776-3c10b7379d9c-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-2shrx\" (UID: \"7c4da269-7f3c-4101-8776-3c10b7379d9c\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.939180 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwd7c\" (UniqueName: \"kubernetes.io/projected/931ee02f-2887-4831-a0df-94eccdf81672-kube-api-access-pwd7c\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.946844 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsgf9\" (UniqueName: \"kubernetes.io/projected/f2005529-2d5f-4dc3-b783-d30e0e1d7962-kube-api-access-wsgf9\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.949395 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzxqw\" (UniqueName: \"kubernetes.io/projected/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-kube-api-access-xzxqw\") pod \"controller-5bddd4b946-55pxp\" (UID: \"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29\") " pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.949937 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsqx7\" (UniqueName: \"kubernetes.io/projected/7c4da269-7f3c-4101-8776-3c10b7379d9c-kube-api-access-hsqx7\") pod \"frr-k8s-webhook-server-7784b6fcf-2shrx\" (UID: \"7c4da269-7f3c-4101-8776-3c10b7379d9c\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" Dec 13 17:28:48 crc kubenswrapper[4989]: I1213 17:28:48.986372 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.239060 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx"] Dec 13 17:28:49 crc kubenswrapper[4989]: W1213 17:28:49.242484 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c4da269_7f3c_4101_8776_3c10b7379d9c.slice/crio-2e92c265366bf06a3dbe4cca8dc2e839e4f7b413f3e2f0943f5fb21d610ebeaa WatchSource:0}: Error finding container 2e92c265366bf06a3dbe4cca8dc2e839e4f7b413f3e2f0943f5fb21d610ebeaa: Status 404 returned error can't find the container with id 2e92c265366bf06a3dbe4cca8dc2e839e4f7b413f3e2f0943f5fb21d610ebeaa Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.420544 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2005529-2d5f-4dc3-b783-d30e0e1d7962-metrics-certs\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.420593 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-metrics-certs\") pod \"controller-5bddd4b946-55pxp\" (UID: \"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29\") " pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.420635 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-memberlist\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.420659 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-metrics-certs\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:49 crc kubenswrapper[4989]: E1213 17:28:49.421652 4989 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 13 17:28:49 crc kubenswrapper[4989]: E1213 17:28:49.421726 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-memberlist podName:931ee02f-2887-4831-a0df-94eccdf81672 nodeName:}" failed. No retries permitted until 2025-12-13 17:28:50.421710312 +0000 UTC m=+745.028157450 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-memberlist") pod "speaker-r5qp2" (UID: "931ee02f-2887-4831-a0df-94eccdf81672") : secret "metallb-memberlist" not found Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.427971 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29-metrics-certs\") pod \"controller-5bddd4b946-55pxp\" (UID: \"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29\") " pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.428739 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-metrics-certs\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.429773 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f2005529-2d5f-4dc3-b783-d30e0e1d7962-metrics-certs\") pod \"frr-k8s-hq8z9\" (UID: \"f2005529-2d5f-4dc3-b783-d30e0e1d7962\") " pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.573781 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.672258 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.864062 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-55pxp"] Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.948435 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hq8z9" event={"ID":"f2005529-2d5f-4dc3-b783-d30e0e1d7962","Type":"ContainerStarted","Data":"8f8ab28839c73c276092907d24ad2dbd9fbc3fe488765b68561896c0c7e84644"} Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.950397 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" event={"ID":"7c4da269-7f3c-4101-8776-3c10b7379d9c","Type":"ContainerStarted","Data":"2e92c265366bf06a3dbe4cca8dc2e839e4f7b413f3e2f0943f5fb21d610ebeaa"} Dec 13 17:28:49 crc kubenswrapper[4989]: I1213 17:28:49.951664 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-55pxp" event={"ID":"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29","Type":"ContainerStarted","Data":"c41a2cf1b3487ad31aeedc2141e32239a156edc58d5bcef938f547e823ef793b"} Dec 13 17:28:50 crc kubenswrapper[4989]: I1213 17:28:50.435342 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-memberlist\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:50 crc kubenswrapper[4989]: I1213 17:28:50.453185 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/931ee02f-2887-4831-a0df-94eccdf81672-memberlist\") pod \"speaker-r5qp2\" (UID: \"931ee02f-2887-4831-a0df-94eccdf81672\") " pod="metallb-system/speaker-r5qp2" Dec 13 17:28:50 crc kubenswrapper[4989]: I1213 17:28:50.554379 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-r5qp2" Dec 13 17:28:50 crc kubenswrapper[4989]: I1213 17:28:50.974364 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r5qp2" event={"ID":"931ee02f-2887-4831-a0df-94eccdf81672","Type":"ContainerStarted","Data":"5c3db3bd12e639206413f2c11d3e057a26acbf809fef29ca4f2917c273c2e27a"} Dec 13 17:28:50 crc kubenswrapper[4989]: I1213 17:28:50.974408 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r5qp2" event={"ID":"931ee02f-2887-4831-a0df-94eccdf81672","Type":"ContainerStarted","Data":"cab91455af26ad216f9f5ea6d080820cb06f60af8cef385cdbc9e844b95ad4e3"} Dec 13 17:28:50 crc kubenswrapper[4989]: I1213 17:28:50.976909 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-55pxp" event={"ID":"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29","Type":"ContainerStarted","Data":"9f24db04778fa1ba59b64d09f2888d4ae1f6f37cb5e55e4a42eab1c659e5fa91"} Dec 13 17:28:54 crc kubenswrapper[4989]: I1213 17:28:54.000534 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-r5qp2" event={"ID":"931ee02f-2887-4831-a0df-94eccdf81672","Type":"ContainerStarted","Data":"15ba9e1de19e6d38039c9aac84a121c6f97c1da1504cb0a540a93c0e6cc00a1c"} Dec 13 17:28:54 crc kubenswrapper[4989]: I1213 17:28:54.002919 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-r5qp2" Dec 13 17:28:54 crc kubenswrapper[4989]: I1213 17:28:54.006815 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-55pxp" event={"ID":"c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29","Type":"ContainerStarted","Data":"15d2e36a9dc7288c13779464fe48ab137801155fbe2268919f9bd838958bca7b"} Dec 13 17:28:54 crc kubenswrapper[4989]: I1213 17:28:54.006990 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:28:54 crc kubenswrapper[4989]: I1213 17:28:54.024918 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-r5qp2" podStartSLOduration=3.212795431 podStartE2EDuration="6.024897353s" podCreationTimestamp="2025-12-13 17:28:48 +0000 UTC" firstStartedPulling="2025-12-13 17:28:50.898981341 +0000 UTC m=+745.505428479" lastFinishedPulling="2025-12-13 17:28:53.711083263 +0000 UTC m=+748.317530401" observedRunningTime="2025-12-13 17:28:54.021295342 +0000 UTC m=+748.627742480" watchObservedRunningTime="2025-12-13 17:28:54.024897353 +0000 UTC m=+748.631344501" Dec 13 17:28:54 crc kubenswrapper[4989]: I1213 17:28:54.051221 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5bddd4b946-55pxp" podStartSLOduration=2.403158969 podStartE2EDuration="6.051177838s" podCreationTimestamp="2025-12-13 17:28:48 +0000 UTC" firstStartedPulling="2025-12-13 17:28:50.033111637 +0000 UTC m=+744.639558775" lastFinishedPulling="2025-12-13 17:28:53.681130506 +0000 UTC m=+748.287577644" observedRunningTime="2025-12-13 17:28:54.048184064 +0000 UTC m=+748.654631192" watchObservedRunningTime="2025-12-13 17:28:54.051177838 +0000 UTC m=+748.657624986" Dec 13 17:28:59 crc kubenswrapper[4989]: I1213 17:28:59.044958 4989 generic.go:334] "Generic (PLEG): container finished" podID="f2005529-2d5f-4dc3-b783-d30e0e1d7962" containerID="e631da446705d5a2d006d522a2881d043fdfdfb457d8434f01ca6dfae6fd233d" exitCode=0 Dec 13 17:28:59 crc kubenswrapper[4989]: I1213 17:28:59.045027 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hq8z9" event={"ID":"f2005529-2d5f-4dc3-b783-d30e0e1d7962","Type":"ContainerDied","Data":"e631da446705d5a2d006d522a2881d043fdfdfb457d8434f01ca6dfae6fd233d"} Dec 13 17:28:59 crc kubenswrapper[4989]: I1213 17:28:59.048019 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" event={"ID":"7c4da269-7f3c-4101-8776-3c10b7379d9c","Type":"ContainerStarted","Data":"2ef3efda3b7e4d2e1da832e4e4d3756cf7f3de487af7db340fb88de06ed3d93c"} Dec 13 17:28:59 crc kubenswrapper[4989]: I1213 17:28:59.048213 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" Dec 13 17:28:59 crc kubenswrapper[4989]: I1213 17:28:59.125174 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" podStartSLOduration=2.454459665 podStartE2EDuration="11.125151429s" podCreationTimestamp="2025-12-13 17:28:48 +0000 UTC" firstStartedPulling="2025-12-13 17:28:49.245934595 +0000 UTC m=+743.852381733" lastFinishedPulling="2025-12-13 17:28:57.916626359 +0000 UTC m=+752.523073497" observedRunningTime="2025-12-13 17:28:59.120662304 +0000 UTC m=+753.727109462" watchObservedRunningTime="2025-12-13 17:28:59.125151429 +0000 UTC m=+753.731598577" Dec 13 17:29:00 crc kubenswrapper[4989]: I1213 17:29:00.057145 4989 generic.go:334] "Generic (PLEG): container finished" podID="f2005529-2d5f-4dc3-b783-d30e0e1d7962" containerID="6fd77cf7aa314a3d9f26a7f7c755caeb97f604d0f89655f020362c54733aa84d" exitCode=0 Dec 13 17:29:00 crc kubenswrapper[4989]: I1213 17:29:00.057205 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hq8z9" event={"ID":"f2005529-2d5f-4dc3-b783-d30e0e1d7962","Type":"ContainerDied","Data":"6fd77cf7aa314a3d9f26a7f7c755caeb97f604d0f89655f020362c54733aa84d"} Dec 13 17:29:00 crc kubenswrapper[4989]: I1213 17:29:00.559345 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-r5qp2" Dec 13 17:29:01 crc kubenswrapper[4989]: I1213 17:29:01.063598 4989 generic.go:334] "Generic (PLEG): container finished" podID="f2005529-2d5f-4dc3-b783-d30e0e1d7962" containerID="aa1bb3f84464ac3e2ab4fabb75b9d19cad0710631c94563178c5a823f882263a" exitCode=0 Dec 13 17:29:01 crc kubenswrapper[4989]: I1213 17:29:01.063631 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hq8z9" event={"ID":"f2005529-2d5f-4dc3-b783-d30e0e1d7962","Type":"ContainerDied","Data":"aa1bb3f84464ac3e2ab4fabb75b9d19cad0710631c94563178c5a823f882263a"} Dec 13 17:29:02 crc kubenswrapper[4989]: I1213 17:29:02.081519 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hq8z9" event={"ID":"f2005529-2d5f-4dc3-b783-d30e0e1d7962","Type":"ContainerStarted","Data":"8c6e5119430bf3d328950e1b1a382f7f900592d0ce59e976e1a47ea15c428840"} Dec 13 17:29:02 crc kubenswrapper[4989]: I1213 17:29:02.081559 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hq8z9" event={"ID":"f2005529-2d5f-4dc3-b783-d30e0e1d7962","Type":"ContainerStarted","Data":"72a104490aada7b469b0297ff2dbe1c32de23dc15c5ebd4e437a8d67a586e8b2"} Dec 13 17:29:02 crc kubenswrapper[4989]: I1213 17:29:02.081570 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hq8z9" event={"ID":"f2005529-2d5f-4dc3-b783-d30e0e1d7962","Type":"ContainerStarted","Data":"37a31581926537877e175941ba05813ebf28bdb91362b6866934d058dc3cfad1"} Dec 13 17:29:02 crc kubenswrapper[4989]: I1213 17:29:02.081578 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hq8z9" event={"ID":"f2005529-2d5f-4dc3-b783-d30e0e1d7962","Type":"ContainerStarted","Data":"a0bd404f7ef10ed18ac038a15088d1fa413b75e7ab21f43105009432bd0c5301"} Dec 13 17:29:02 crc kubenswrapper[4989]: I1213 17:29:02.081586 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hq8z9" event={"ID":"f2005529-2d5f-4dc3-b783-d30e0e1d7962","Type":"ContainerStarted","Data":"68f7d575d84635fe01fb08d5cf785996abb8f422ac6abcb691cb589671c412f8"} Dec 13 17:29:02 crc kubenswrapper[4989]: I1213 17:29:02.081594 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hq8z9" event={"ID":"f2005529-2d5f-4dc3-b783-d30e0e1d7962","Type":"ContainerStarted","Data":"cdc2510586f460ba80fed6edf1909cfc71ead41f268746e0a4ebde564e4a0e07"} Dec 13 17:29:02 crc kubenswrapper[4989]: I1213 17:29:02.081762 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:29:02 crc kubenswrapper[4989]: I1213 17:29:02.652138 4989 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 13 17:29:04 crc kubenswrapper[4989]: I1213 17:29:04.574614 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:29:04 crc kubenswrapper[4989]: I1213 17:29:04.620750 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:29:04 crc kubenswrapper[4989]: I1213 17:29:04.647971 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hq8z9" podStartSLOduration=8.433549748 podStartE2EDuration="16.647951687s" podCreationTimestamp="2025-12-13 17:28:48 +0000 UTC" firstStartedPulling="2025-12-13 17:28:49.731100078 +0000 UTC m=+744.337547216" lastFinishedPulling="2025-12-13 17:28:57.945502017 +0000 UTC m=+752.551949155" observedRunningTime="2025-12-13 17:29:02.112256117 +0000 UTC m=+756.718703315" watchObservedRunningTime="2025-12-13 17:29:04.647951687 +0000 UTC m=+759.254398835" Dec 13 17:29:06 crc kubenswrapper[4989]: I1213 17:29:06.748839 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-cmjdc"] Dec 13 17:29:06 crc kubenswrapper[4989]: I1213 17:29:06.749646 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-cmjdc" Dec 13 17:29:06 crc kubenswrapper[4989]: I1213 17:29:06.752344 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 13 17:29:06 crc kubenswrapper[4989]: I1213 17:29:06.752522 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 13 17:29:06 crc kubenswrapper[4989]: I1213 17:29:06.752706 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-gwpcs" Dec 13 17:29:06 crc kubenswrapper[4989]: I1213 17:29:06.784563 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlr42\" (UniqueName: \"kubernetes.io/projected/6fec4527-4426-4752-a913-20781b3df09d-kube-api-access-hlr42\") pod \"mariadb-operator-index-cmjdc\" (UID: \"6fec4527-4426-4752-a913-20781b3df09d\") " pod="openstack-operators/mariadb-operator-index-cmjdc" Dec 13 17:29:06 crc kubenswrapper[4989]: I1213 17:29:06.790743 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-cmjdc"] Dec 13 17:29:06 crc kubenswrapper[4989]: I1213 17:29:06.888520 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlr42\" (UniqueName: \"kubernetes.io/projected/6fec4527-4426-4752-a913-20781b3df09d-kube-api-access-hlr42\") pod \"mariadb-operator-index-cmjdc\" (UID: \"6fec4527-4426-4752-a913-20781b3df09d\") " pod="openstack-operators/mariadb-operator-index-cmjdc" Dec 13 17:29:06 crc kubenswrapper[4989]: I1213 17:29:06.911874 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlr42\" (UniqueName: \"kubernetes.io/projected/6fec4527-4426-4752-a913-20781b3df09d-kube-api-access-hlr42\") pod \"mariadb-operator-index-cmjdc\" (UID: \"6fec4527-4426-4752-a913-20781b3df09d\") " pod="openstack-operators/mariadb-operator-index-cmjdc" Dec 13 17:29:07 crc kubenswrapper[4989]: I1213 17:29:07.082438 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-cmjdc" Dec 13 17:29:07 crc kubenswrapper[4989]: I1213 17:29:07.539824 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-cmjdc"] Dec 13 17:29:07 crc kubenswrapper[4989]: W1213 17:29:07.545972 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fec4527_4426_4752_a913_20781b3df09d.slice/crio-29bd27fb76bcc5037e16b13c2ef7ec676040e2c5670f4f5a5edfa5b694d9f537 WatchSource:0}: Error finding container 29bd27fb76bcc5037e16b13c2ef7ec676040e2c5670f4f5a5edfa5b694d9f537: Status 404 returned error can't find the container with id 29bd27fb76bcc5037e16b13c2ef7ec676040e2c5670f4f5a5edfa5b694d9f537 Dec 13 17:29:08 crc kubenswrapper[4989]: I1213 17:29:08.115360 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-cmjdc" event={"ID":"6fec4527-4426-4752-a913-20781b3df09d","Type":"ContainerStarted","Data":"29bd27fb76bcc5037e16b13c2ef7ec676040e2c5670f4f5a5edfa5b694d9f537"} Dec 13 17:29:08 crc kubenswrapper[4989]: I1213 17:29:08.997132 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-2shrx" Dec 13 17:29:09 crc kubenswrapper[4989]: I1213 17:29:09.121831 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-cmjdc" event={"ID":"6fec4527-4426-4752-a913-20781b3df09d","Type":"ContainerStarted","Data":"fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8"} Dec 13 17:29:09 crc kubenswrapper[4989]: I1213 17:29:09.676466 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5bddd4b946-55pxp" Dec 13 17:29:09 crc kubenswrapper[4989]: I1213 17:29:09.695196 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-cmjdc" podStartSLOduration=2.755519131 podStartE2EDuration="3.695173349s" podCreationTimestamp="2025-12-13 17:29:06 +0000 UTC" firstStartedPulling="2025-12-13 17:29:07.547546456 +0000 UTC m=+762.153993594" lastFinishedPulling="2025-12-13 17:29:08.487200674 +0000 UTC m=+763.093647812" observedRunningTime="2025-12-13 17:29:09.137552929 +0000 UTC m=+763.744000087" watchObservedRunningTime="2025-12-13 17:29:09.695173349 +0000 UTC m=+764.301620487" Dec 13 17:29:10 crc kubenswrapper[4989]: I1213 17:29:10.118389 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-cmjdc"] Dec 13 17:29:10 crc kubenswrapper[4989]: I1213 17:29:10.726736 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-zcs27"] Dec 13 17:29:10 crc kubenswrapper[4989]: I1213 17:29:10.727461 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zcs27" Dec 13 17:29:10 crc kubenswrapper[4989]: I1213 17:29:10.741943 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-zcs27"] Dec 13 17:29:10 crc kubenswrapper[4989]: I1213 17:29:10.743889 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5d2d\" (UniqueName: \"kubernetes.io/projected/3a635a9e-69b4-4b82-b9fe-207ac9e5184e-kube-api-access-l5d2d\") pod \"mariadb-operator-index-zcs27\" (UID: \"3a635a9e-69b4-4b82-b9fe-207ac9e5184e\") " pod="openstack-operators/mariadb-operator-index-zcs27" Dec 13 17:29:10 crc kubenswrapper[4989]: I1213 17:29:10.846099 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5d2d\" (UniqueName: \"kubernetes.io/projected/3a635a9e-69b4-4b82-b9fe-207ac9e5184e-kube-api-access-l5d2d\") pod \"mariadb-operator-index-zcs27\" (UID: \"3a635a9e-69b4-4b82-b9fe-207ac9e5184e\") " pod="openstack-operators/mariadb-operator-index-zcs27" Dec 13 17:29:10 crc kubenswrapper[4989]: I1213 17:29:10.864287 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5d2d\" (UniqueName: \"kubernetes.io/projected/3a635a9e-69b4-4b82-b9fe-207ac9e5184e-kube-api-access-l5d2d\") pod \"mariadb-operator-index-zcs27\" (UID: \"3a635a9e-69b4-4b82-b9fe-207ac9e5184e\") " pod="openstack-operators/mariadb-operator-index-zcs27" Dec 13 17:29:11 crc kubenswrapper[4989]: I1213 17:29:11.102538 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zcs27" Dec 13 17:29:11 crc kubenswrapper[4989]: I1213 17:29:11.137250 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-cmjdc" podUID="6fec4527-4426-4752-a913-20781b3df09d" containerName="registry-server" containerID="cri-o://fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8" gracePeriod=2 Dec 13 17:29:11 crc kubenswrapper[4989]: I1213 17:29:11.338431 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-zcs27"] Dec 13 17:29:11 crc kubenswrapper[4989]: W1213 17:29:11.348842 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a635a9e_69b4_4b82_b9fe_207ac9e5184e.slice/crio-db651614d4764b6bc695582240d1a5ac01d3d28c50005f21e7b8fb1a9ec73062 WatchSource:0}: Error finding container db651614d4764b6bc695582240d1a5ac01d3d28c50005f21e7b8fb1a9ec73062: Status 404 returned error can't find the container with id db651614d4764b6bc695582240d1a5ac01d3d28c50005f21e7b8fb1a9ec73062 Dec 13 17:29:11 crc kubenswrapper[4989]: I1213 17:29:11.455948 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-cmjdc" Dec 13 17:29:11 crc kubenswrapper[4989]: I1213 17:29:11.656253 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlr42\" (UniqueName: \"kubernetes.io/projected/6fec4527-4426-4752-a913-20781b3df09d-kube-api-access-hlr42\") pod \"6fec4527-4426-4752-a913-20781b3df09d\" (UID: \"6fec4527-4426-4752-a913-20781b3df09d\") " Dec 13 17:29:11 crc kubenswrapper[4989]: I1213 17:29:11.660479 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fec4527-4426-4752-a913-20781b3df09d-kube-api-access-hlr42" (OuterVolumeSpecName: "kube-api-access-hlr42") pod "6fec4527-4426-4752-a913-20781b3df09d" (UID: "6fec4527-4426-4752-a913-20781b3df09d"). InnerVolumeSpecName "kube-api-access-hlr42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:29:11 crc kubenswrapper[4989]: I1213 17:29:11.758334 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlr42\" (UniqueName: \"kubernetes.io/projected/6fec4527-4426-4752-a913-20781b3df09d-kube-api-access-hlr42\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.144362 4989 generic.go:334] "Generic (PLEG): container finished" podID="6fec4527-4426-4752-a913-20781b3df09d" containerID="fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8" exitCode=0 Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.144429 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-cmjdc" event={"ID":"6fec4527-4426-4752-a913-20781b3df09d","Type":"ContainerDied","Data":"fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8"} Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.144682 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-cmjdc" event={"ID":"6fec4527-4426-4752-a913-20781b3df09d","Type":"ContainerDied","Data":"29bd27fb76bcc5037e16b13c2ef7ec676040e2c5670f4f5a5edfa5b694d9f537"} Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.144704 4989 scope.go:117] "RemoveContainer" containerID="fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8" Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.144478 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-cmjdc" Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.146377 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zcs27" event={"ID":"3a635a9e-69b4-4b82-b9fe-207ac9e5184e","Type":"ContainerStarted","Data":"6ff595da2c13ecdb90d6cdd271e0ce54dcc432b791ae68955cedd7c483848ff0"} Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.146402 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zcs27" event={"ID":"3a635a9e-69b4-4b82-b9fe-207ac9e5184e","Type":"ContainerStarted","Data":"db651614d4764b6bc695582240d1a5ac01d3d28c50005f21e7b8fb1a9ec73062"} Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.158677 4989 scope.go:117] "RemoveContainer" containerID="fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8" Dec 13 17:29:12 crc kubenswrapper[4989]: E1213 17:29:12.159175 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8\": container with ID starting with fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8 not found: ID does not exist" containerID="fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8" Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.159236 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8"} err="failed to get container status \"fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8\": rpc error: code = NotFound desc = could not find container \"fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8\": container with ID starting with fe5ea40f07c795a8315fc981767085dc094db7ed9da50eefb5c8a868f09bfdb8 not found: ID does not exist" Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.161883 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-cmjdc"] Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.165209 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-cmjdc"] Dec 13 17:29:12 crc kubenswrapper[4989]: I1213 17:29:12.175743 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-zcs27" podStartSLOduration=1.653374862 podStartE2EDuration="2.175724806s" podCreationTimestamp="2025-12-13 17:29:10 +0000 UTC" firstStartedPulling="2025-12-13 17:29:11.353562264 +0000 UTC m=+765.960009402" lastFinishedPulling="2025-12-13 17:29:11.875912208 +0000 UTC m=+766.482359346" observedRunningTime="2025-12-13 17:29:12.17376582 +0000 UTC m=+766.780212958" watchObservedRunningTime="2025-12-13 17:29:12.175724806 +0000 UTC m=+766.782171944" Dec 13 17:29:14 crc kubenswrapper[4989]: I1213 17:29:14.020526 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fec4527-4426-4752-a913-20781b3df09d" path="/var/lib/kubelet/pods/6fec4527-4426-4752-a913-20781b3df09d/volumes" Dec 13 17:29:19 crc kubenswrapper[4989]: I1213 17:29:19.578495 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hq8z9" Dec 13 17:29:21 crc kubenswrapper[4989]: I1213 17:29:21.103148 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-zcs27" Dec 13 17:29:21 crc kubenswrapper[4989]: I1213 17:29:21.103203 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-zcs27" Dec 13 17:29:21 crc kubenswrapper[4989]: I1213 17:29:21.134400 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-zcs27" Dec 13 17:29:21 crc kubenswrapper[4989]: I1213 17:29:21.235911 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-zcs27" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.552368 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cxz2g"] Dec 13 17:29:22 crc kubenswrapper[4989]: E1213 17:29:22.552855 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fec4527-4426-4752-a913-20781b3df09d" containerName="registry-server" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.552885 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fec4527-4426-4752-a913-20781b3df09d" containerName="registry-server" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.553085 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fec4527-4426-4752-a913-20781b3df09d" containerName="registry-server" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.554724 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.577530 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cxz2g"] Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.608417 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqrh2\" (UniqueName: \"kubernetes.io/projected/f6c5fbf4-620a-45c5-85cd-04fb5f227247-kube-api-access-vqrh2\") pod \"redhat-operators-cxz2g\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.608487 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-utilities\") pod \"redhat-operators-cxz2g\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.608525 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-catalog-content\") pod \"redhat-operators-cxz2g\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.709557 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqrh2\" (UniqueName: \"kubernetes.io/projected/f6c5fbf4-620a-45c5-85cd-04fb5f227247-kube-api-access-vqrh2\") pod \"redhat-operators-cxz2g\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.709622 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-utilities\") pod \"redhat-operators-cxz2g\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.709653 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-catalog-content\") pod \"redhat-operators-cxz2g\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.710135 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-catalog-content\") pod \"redhat-operators-cxz2g\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.710445 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-utilities\") pod \"redhat-operators-cxz2g\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.736258 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqrh2\" (UniqueName: \"kubernetes.io/projected/f6c5fbf4-620a-45c5-85cd-04fb5f227247-kube-api-access-vqrh2\") pod \"redhat-operators-cxz2g\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.774104 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s"] Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.775171 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.777267 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k762t" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.790161 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s"] Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.897806 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.914248 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-bundle\") pod \"5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.914552 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-util\") pod \"5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:22 crc kubenswrapper[4989]: I1213 17:29:22.914625 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-972j6\" (UniqueName: \"kubernetes.io/projected/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-kube-api-access-972j6\") pod \"5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:23 crc kubenswrapper[4989]: I1213 17:29:23.017542 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-972j6\" (UniqueName: \"kubernetes.io/projected/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-kube-api-access-972j6\") pod \"5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:23 crc kubenswrapper[4989]: I1213 17:29:23.017628 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-bundle\") pod \"5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:23 crc kubenswrapper[4989]: I1213 17:29:23.017659 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-util\") pod \"5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:23 crc kubenswrapper[4989]: I1213 17:29:23.018091 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-util\") pod \"5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:23 crc kubenswrapper[4989]: I1213 17:29:23.018310 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-bundle\") pod \"5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:23 crc kubenswrapper[4989]: I1213 17:29:23.041719 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-972j6\" (UniqueName: \"kubernetes.io/projected/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-kube-api-access-972j6\") pod \"5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:23 crc kubenswrapper[4989]: I1213 17:29:23.091960 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:23 crc kubenswrapper[4989]: I1213 17:29:23.114037 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cxz2g"] Dec 13 17:29:23 crc kubenswrapper[4989]: W1213 17:29:23.123761 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6c5fbf4_620a_45c5_85cd_04fb5f227247.slice/crio-80f8c7e9356d9e8bd4dc108380120ed82417f19cb663c5dd3fd9b6818b86a531 WatchSource:0}: Error finding container 80f8c7e9356d9e8bd4dc108380120ed82417f19cb663c5dd3fd9b6818b86a531: Status 404 returned error can't find the container with id 80f8c7e9356d9e8bd4dc108380120ed82417f19cb663c5dd3fd9b6818b86a531 Dec 13 17:29:23 crc kubenswrapper[4989]: I1213 17:29:23.235826 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxz2g" event={"ID":"f6c5fbf4-620a-45c5-85cd-04fb5f227247","Type":"ContainerStarted","Data":"80f8c7e9356d9e8bd4dc108380120ed82417f19cb663c5dd3fd9b6818b86a531"} Dec 13 17:29:23 crc kubenswrapper[4989]: I1213 17:29:23.318344 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s"] Dec 13 17:29:23 crc kubenswrapper[4989]: W1213 17:29:23.327613 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fa079b8_c658_4ba8_b0a4_0034c70e2d7d.slice/crio-e53963ca453858e87e74de65c8e56330571eca852ee14c63340828614800187d WatchSource:0}: Error finding container e53963ca453858e87e74de65c8e56330571eca852ee14c63340828614800187d: Status 404 returned error can't find the container with id e53963ca453858e87e74de65c8e56330571eca852ee14c63340828614800187d Dec 13 17:29:24 crc kubenswrapper[4989]: I1213 17:29:24.253748 4989 generic.go:334] "Generic (PLEG): container finished" podID="5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" containerID="840ecc0ff6c76d7c717a8a2c4f4c54600b99f2864110036f4a3e12e55ed827f3" exitCode=0 Dec 13 17:29:24 crc kubenswrapper[4989]: I1213 17:29:24.253856 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" event={"ID":"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d","Type":"ContainerDied","Data":"840ecc0ff6c76d7c717a8a2c4f4c54600b99f2864110036f4a3e12e55ed827f3"} Dec 13 17:29:24 crc kubenswrapper[4989]: I1213 17:29:24.255832 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" event={"ID":"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d","Type":"ContainerStarted","Data":"e53963ca453858e87e74de65c8e56330571eca852ee14c63340828614800187d"} Dec 13 17:29:24 crc kubenswrapper[4989]: I1213 17:29:24.257693 4989 generic.go:334] "Generic (PLEG): container finished" podID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerID="9c0c24711280c50a6caab7bb32276e7bbebb864a1f8561c7a1e31e8242f4701a" exitCode=0 Dec 13 17:29:24 crc kubenswrapper[4989]: I1213 17:29:24.257733 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxz2g" event={"ID":"f6c5fbf4-620a-45c5-85cd-04fb5f227247","Type":"ContainerDied","Data":"9c0c24711280c50a6caab7bb32276e7bbebb864a1f8561c7a1e31e8242f4701a"} Dec 13 17:29:27 crc kubenswrapper[4989]: I1213 17:29:27.275629 4989 generic.go:334] "Generic (PLEG): container finished" podID="5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" containerID="66a9891c3b5ca710046e7f8b45890b09bf2086d4cb2cb1a6c247dfeb3a55a95d" exitCode=0 Dec 13 17:29:27 crc kubenswrapper[4989]: I1213 17:29:27.275705 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" event={"ID":"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d","Type":"ContainerDied","Data":"66a9891c3b5ca710046e7f8b45890b09bf2086d4cb2cb1a6c247dfeb3a55a95d"} Dec 13 17:29:27 crc kubenswrapper[4989]: I1213 17:29:27.280417 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxz2g" event={"ID":"f6c5fbf4-620a-45c5-85cd-04fb5f227247","Type":"ContainerStarted","Data":"f5338fe03fe28709ebcc8d9503949d7a190f4e378b500c397caa9ac940a35146"} Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.293247 4989 generic.go:334] "Generic (PLEG): container finished" podID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerID="f5338fe03fe28709ebcc8d9503949d7a190f4e378b500c397caa9ac940a35146" exitCode=0 Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.293354 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxz2g" event={"ID":"f6c5fbf4-620a-45c5-85cd-04fb5f227247","Type":"ContainerDied","Data":"f5338fe03fe28709ebcc8d9503949d7a190f4e378b500c397caa9ac940a35146"} Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.297115 4989 generic.go:334] "Generic (PLEG): container finished" podID="5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" containerID="b2e20b8dc38bf625f7e62af2d3e368be4afc7fbd9cbe17074aa5612c72a3d2f2" exitCode=0 Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.297155 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" event={"ID":"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d","Type":"ContainerDied","Data":"b2e20b8dc38bf625f7e62af2d3e368be4afc7fbd9cbe17074aa5612c72a3d2f2"} Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.527178 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hqxjt"] Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.528386 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.547625 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hqxjt"] Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.602371 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-utilities\") pod \"certified-operators-hqxjt\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.602447 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-catalog-content\") pod \"certified-operators-hqxjt\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.602483 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txsmv\" (UniqueName: \"kubernetes.io/projected/f75af986-e7f9-48e9-8680-eb09e74cd9fb-kube-api-access-txsmv\") pod \"certified-operators-hqxjt\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.703934 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txsmv\" (UniqueName: \"kubernetes.io/projected/f75af986-e7f9-48e9-8680-eb09e74cd9fb-kube-api-access-txsmv\") pod \"certified-operators-hqxjt\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.703991 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-utilities\") pod \"certified-operators-hqxjt\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.704040 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-catalog-content\") pod \"certified-operators-hqxjt\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.704458 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-catalog-content\") pod \"certified-operators-hqxjt\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.704584 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-utilities\") pod \"certified-operators-hqxjt\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.728195 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txsmv\" (UniqueName: \"kubernetes.io/projected/f75af986-e7f9-48e9-8680-eb09e74cd9fb-kube-api-access-txsmv\") pod \"certified-operators-hqxjt\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:28 crc kubenswrapper[4989]: I1213 17:29:28.843267 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:29 crc kubenswrapper[4989]: I1213 17:29:29.392353 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hqxjt"] Dec 13 17:29:29 crc kubenswrapper[4989]: I1213 17:29:29.640399 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:29 crc kubenswrapper[4989]: I1213 17:29:29.646875 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-bundle\") pod \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " Dec 13 17:29:29 crc kubenswrapper[4989]: I1213 17:29:29.646948 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-972j6\" (UniqueName: \"kubernetes.io/projected/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-kube-api-access-972j6\") pod \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " Dec 13 17:29:29 crc kubenswrapper[4989]: I1213 17:29:29.646974 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-util\") pod \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\" (UID: \"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d\") " Dec 13 17:29:29 crc kubenswrapper[4989]: I1213 17:29:29.647977 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-bundle" (OuterVolumeSpecName: "bundle") pod "5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" (UID: "5fa079b8-c658-4ba8-b0a4-0034c70e2d7d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:29:29 crc kubenswrapper[4989]: I1213 17:29:29.651995 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-kube-api-access-972j6" (OuterVolumeSpecName: "kube-api-access-972j6") pod "5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" (UID: "5fa079b8-c658-4ba8-b0a4-0034c70e2d7d"). InnerVolumeSpecName "kube-api-access-972j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:29:29 crc kubenswrapper[4989]: I1213 17:29:29.658138 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-util" (OuterVolumeSpecName: "util") pod "5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" (UID: "5fa079b8-c658-4ba8-b0a4-0034c70e2d7d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:29:29 crc kubenswrapper[4989]: I1213 17:29:29.748649 4989 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:29 crc kubenswrapper[4989]: I1213 17:29:29.748996 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-972j6\" (UniqueName: \"kubernetes.io/projected/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-kube-api-access-972j6\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:29 crc kubenswrapper[4989]: I1213 17:29:29.749009 4989 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d-util\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:30 crc kubenswrapper[4989]: I1213 17:29:30.313000 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" event={"ID":"5fa079b8-c658-4ba8-b0a4-0034c70e2d7d","Type":"ContainerDied","Data":"e53963ca453858e87e74de65c8e56330571eca852ee14c63340828614800187d"} Dec 13 17:29:30 crc kubenswrapper[4989]: I1213 17:29:30.313039 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s" Dec 13 17:29:30 crc kubenswrapper[4989]: I1213 17:29:30.313050 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e53963ca453858e87e74de65c8e56330571eca852ee14c63340828614800187d" Dec 13 17:29:30 crc kubenswrapper[4989]: I1213 17:29:30.314608 4989 generic.go:334] "Generic (PLEG): container finished" podID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" containerID="e616b212936318cebfc0623337103dd737b227be59f40c43fda4c160f806f73a" exitCode=0 Dec 13 17:29:30 crc kubenswrapper[4989]: I1213 17:29:30.314682 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxjt" event={"ID":"f75af986-e7f9-48e9-8680-eb09e74cd9fb","Type":"ContainerDied","Data":"e616b212936318cebfc0623337103dd737b227be59f40c43fda4c160f806f73a"} Dec 13 17:29:30 crc kubenswrapper[4989]: I1213 17:29:30.314706 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxjt" event={"ID":"f75af986-e7f9-48e9-8680-eb09e74cd9fb","Type":"ContainerStarted","Data":"8a1f6648b4f354c46c48d6100e69c8f379f39d671c878e50dbd2abf3547443f2"} Dec 13 17:29:30 crc kubenswrapper[4989]: I1213 17:29:30.317246 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxz2g" event={"ID":"f6c5fbf4-620a-45c5-85cd-04fb5f227247","Type":"ContainerStarted","Data":"96d6c2d24ec8588d0e62191e85b687e0b22d4d6cf3bc58b6c3632286f72c51fe"} Dec 13 17:29:30 crc kubenswrapper[4989]: I1213 17:29:30.359879 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cxz2g" podStartSLOduration=3.498337722 podStartE2EDuration="8.359857419s" podCreationTimestamp="2025-12-13 17:29:22 +0000 UTC" firstStartedPulling="2025-12-13 17:29:24.261407455 +0000 UTC m=+778.867854593" lastFinishedPulling="2025-12-13 17:29:29.122927152 +0000 UTC m=+783.729374290" observedRunningTime="2025-12-13 17:29:30.356780992 +0000 UTC m=+784.963228140" watchObservedRunningTime="2025-12-13 17:29:30.359857419 +0000 UTC m=+784.966304567" Dec 13 17:29:31 crc kubenswrapper[4989]: I1213 17:29:31.323564 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxjt" event={"ID":"f75af986-e7f9-48e9-8680-eb09e74cd9fb","Type":"ContainerStarted","Data":"fa9938b36b3e27a2d80644c38da7dd3d05ab6b023311f59fedd463bd18652212"} Dec 13 17:29:32 crc kubenswrapper[4989]: I1213 17:29:32.331720 4989 generic.go:334] "Generic (PLEG): container finished" podID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" containerID="fa9938b36b3e27a2d80644c38da7dd3d05ab6b023311f59fedd463bd18652212" exitCode=0 Dec 13 17:29:32 crc kubenswrapper[4989]: I1213 17:29:32.331812 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxjt" event={"ID":"f75af986-e7f9-48e9-8680-eb09e74cd9fb","Type":"ContainerDied","Data":"fa9938b36b3e27a2d80644c38da7dd3d05ab6b023311f59fedd463bd18652212"} Dec 13 17:29:32 crc kubenswrapper[4989]: I1213 17:29:32.898325 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:32 crc kubenswrapper[4989]: I1213 17:29:32.898381 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:33 crc kubenswrapper[4989]: I1213 17:29:33.997218 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cxz2g" podUID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerName="registry-server" probeResult="failure" output=< Dec 13 17:29:33 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Dec 13 17:29:33 crc kubenswrapper[4989]: > Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.343038 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxjt" event={"ID":"f75af986-e7f9-48e9-8680-eb09e74cd9fb","Type":"ContainerStarted","Data":"85ba483443ea227de3586c42b7900ddef8a75d1d6f060b12b7da7f5bf2fc13a5"} Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.360430 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hqxjt" podStartSLOduration=3.334861662 podStartE2EDuration="6.360414716s" podCreationTimestamp="2025-12-13 17:29:28 +0000 UTC" firstStartedPulling="2025-12-13 17:29:30.31712919 +0000 UTC m=+784.923576338" lastFinishedPulling="2025-12-13 17:29:33.342682254 +0000 UTC m=+787.949129392" observedRunningTime="2025-12-13 17:29:34.359518598 +0000 UTC m=+788.965965736" watchObservedRunningTime="2025-12-13 17:29:34.360414716 +0000 UTC m=+788.966861854" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.891162 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44"] Dec 13 17:29:34 crc kubenswrapper[4989]: E1213 17:29:34.891430 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" containerName="pull" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.891453 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" containerName="pull" Dec 13 17:29:34 crc kubenswrapper[4989]: E1213 17:29:34.891478 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" containerName="extract" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.891486 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" containerName="extract" Dec 13 17:29:34 crc kubenswrapper[4989]: E1213 17:29:34.891503 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" containerName="util" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.891511 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" containerName="util" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.891634 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" containerName="extract" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.892134 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.894836 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.895146 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qptlx" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.895689 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.903323 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44"] Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.956092 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-apiservice-cert\") pod \"mariadb-operator-controller-manager-57575d5b7f-sxp44\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.956168 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-webhook-cert\") pod \"mariadb-operator-controller-manager-57575d5b7f-sxp44\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:34 crc kubenswrapper[4989]: I1213 17:29:34.956244 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj9kb\" (UniqueName: \"kubernetes.io/projected/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-kube-api-access-rj9kb\") pod \"mariadb-operator-controller-manager-57575d5b7f-sxp44\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:35 crc kubenswrapper[4989]: I1213 17:29:35.057094 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-apiservice-cert\") pod \"mariadb-operator-controller-manager-57575d5b7f-sxp44\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:35 crc kubenswrapper[4989]: I1213 17:29:35.057168 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-webhook-cert\") pod \"mariadb-operator-controller-manager-57575d5b7f-sxp44\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:35 crc kubenswrapper[4989]: I1213 17:29:35.057217 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj9kb\" (UniqueName: \"kubernetes.io/projected/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-kube-api-access-rj9kb\") pod \"mariadb-operator-controller-manager-57575d5b7f-sxp44\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:35 crc kubenswrapper[4989]: I1213 17:29:35.065858 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-apiservice-cert\") pod \"mariadb-operator-controller-manager-57575d5b7f-sxp44\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:35 crc kubenswrapper[4989]: I1213 17:29:35.077345 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-webhook-cert\") pod \"mariadb-operator-controller-manager-57575d5b7f-sxp44\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:35 crc kubenswrapper[4989]: I1213 17:29:35.080670 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj9kb\" (UniqueName: \"kubernetes.io/projected/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-kube-api-access-rj9kb\") pod \"mariadb-operator-controller-manager-57575d5b7f-sxp44\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:35 crc kubenswrapper[4989]: I1213 17:29:35.209843 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:35 crc kubenswrapper[4989]: I1213 17:29:35.507974 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44"] Dec 13 17:29:35 crc kubenswrapper[4989]: W1213 17:29:35.510905 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1fdf4ee_0738_4bb6_a4e6_2f460ed5530f.slice/crio-3e30ad3d228b29de0c5a37d1d836a13cd6016bccb1d2ca1be8c5c167bc4f8bda WatchSource:0}: Error finding container 3e30ad3d228b29de0c5a37d1d836a13cd6016bccb1d2ca1be8c5c167bc4f8bda: Status 404 returned error can't find the container with id 3e30ad3d228b29de0c5a37d1d836a13cd6016bccb1d2ca1be8c5c167bc4f8bda Dec 13 17:29:36 crc kubenswrapper[4989]: I1213 17:29:36.366675 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" event={"ID":"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f","Type":"ContainerStarted","Data":"3e30ad3d228b29de0c5a37d1d836a13cd6016bccb1d2ca1be8c5c167bc4f8bda"} Dec 13 17:29:38 crc kubenswrapper[4989]: I1213 17:29:38.831946 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:29:38 crc kubenswrapper[4989]: I1213 17:29:38.832339 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:29:38 crc kubenswrapper[4989]: I1213 17:29:38.844024 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:38 crc kubenswrapper[4989]: I1213 17:29:38.844075 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:38 crc kubenswrapper[4989]: I1213 17:29:38.900541 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.327530 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xbw5q"] Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.329018 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.341490 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xbw5q"] Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.415210 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-utilities\") pod \"community-operators-xbw5q\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.415273 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-catalog-content\") pod \"community-operators-xbw5q\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.415374 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pwqp\" (UniqueName: \"kubernetes.io/projected/3bba6be3-7805-41e9-a2fb-d50fabe8b254-kube-api-access-8pwqp\") pod \"community-operators-xbw5q\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.479397 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.516194 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pwqp\" (UniqueName: \"kubernetes.io/projected/3bba6be3-7805-41e9-a2fb-d50fabe8b254-kube-api-access-8pwqp\") pod \"community-operators-xbw5q\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.516280 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-utilities\") pod \"community-operators-xbw5q\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.516304 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-catalog-content\") pod \"community-operators-xbw5q\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.516874 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-catalog-content\") pod \"community-operators-xbw5q\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.516885 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-utilities\") pod \"community-operators-xbw5q\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.536099 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pwqp\" (UniqueName: \"kubernetes.io/projected/3bba6be3-7805-41e9-a2fb-d50fabe8b254-kube-api-access-8pwqp\") pod \"community-operators-xbw5q\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:39 crc kubenswrapper[4989]: I1213 17:29:39.656498 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:40 crc kubenswrapper[4989]: I1213 17:29:40.523292 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xbw5q"] Dec 13 17:29:40 crc kubenswrapper[4989]: W1213 17:29:40.529359 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bba6be3_7805_41e9_a2fb_d50fabe8b254.slice/crio-e8a9dd9ad772b44197bbd8be8885dd2abc5f0ad005e7a57020ba67e16b3e7020 WatchSource:0}: Error finding container e8a9dd9ad772b44197bbd8be8885dd2abc5f0ad005e7a57020ba67e16b3e7020: Status 404 returned error can't find the container with id e8a9dd9ad772b44197bbd8be8885dd2abc5f0ad005e7a57020ba67e16b3e7020 Dec 13 17:29:41 crc kubenswrapper[4989]: I1213 17:29:41.419298 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" event={"ID":"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f","Type":"ContainerStarted","Data":"4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94"} Dec 13 17:29:41 crc kubenswrapper[4989]: I1213 17:29:41.419574 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:41 crc kubenswrapper[4989]: I1213 17:29:41.420993 4989 generic.go:334] "Generic (PLEG): container finished" podID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" containerID="56d4e750ad4a4e8bdd118ce322869a0fb4bafb1a369754fa4c515f341b90fb5a" exitCode=0 Dec 13 17:29:41 crc kubenswrapper[4989]: I1213 17:29:41.421035 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xbw5q" event={"ID":"3bba6be3-7805-41e9-a2fb-d50fabe8b254","Type":"ContainerDied","Data":"56d4e750ad4a4e8bdd118ce322869a0fb4bafb1a369754fa4c515f341b90fb5a"} Dec 13 17:29:41 crc kubenswrapper[4989]: I1213 17:29:41.421072 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xbw5q" event={"ID":"3bba6be3-7805-41e9-a2fb-d50fabe8b254","Type":"ContainerStarted","Data":"e8a9dd9ad772b44197bbd8be8885dd2abc5f0ad005e7a57020ba67e16b3e7020"} Dec 13 17:29:41 crc kubenswrapper[4989]: I1213 17:29:41.445699 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" podStartSLOduration=2.617425411 podStartE2EDuration="7.445682127s" podCreationTimestamp="2025-12-13 17:29:34 +0000 UTC" firstStartedPulling="2025-12-13 17:29:35.513244677 +0000 UTC m=+790.119691815" lastFinishedPulling="2025-12-13 17:29:40.341501393 +0000 UTC m=+794.947948531" observedRunningTime="2025-12-13 17:29:41.443467216 +0000 UTC m=+796.049914354" watchObservedRunningTime="2025-12-13 17:29:41.445682127 +0000 UTC m=+796.052129265" Dec 13 17:29:42 crc kubenswrapper[4989]: I1213 17:29:42.519183 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hqxjt"] Dec 13 17:29:42 crc kubenswrapper[4989]: I1213 17:29:42.519415 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hqxjt" podUID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" containerName="registry-server" containerID="cri-o://85ba483443ea227de3586c42b7900ddef8a75d1d6f060b12b7da7f5bf2fc13a5" gracePeriod=2 Dec 13 17:29:42 crc kubenswrapper[4989]: I1213 17:29:42.934971 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:42 crc kubenswrapper[4989]: I1213 17:29:42.971462 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.433423 4989 generic.go:334] "Generic (PLEG): container finished" podID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" containerID="85ba483443ea227de3586c42b7900ddef8a75d1d6f060b12b7da7f5bf2fc13a5" exitCode=0 Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.433601 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxjt" event={"ID":"f75af986-e7f9-48e9-8680-eb09e74cd9fb","Type":"ContainerDied","Data":"85ba483443ea227de3586c42b7900ddef8a75d1d6f060b12b7da7f5bf2fc13a5"} Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.433718 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxjt" event={"ID":"f75af986-e7f9-48e9-8680-eb09e74cd9fb","Type":"ContainerDied","Data":"8a1f6648b4f354c46c48d6100e69c8f379f39d671c878e50dbd2abf3547443f2"} Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.433729 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a1f6648b4f354c46c48d6100e69c8f379f39d671c878e50dbd2abf3547443f2" Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.435607 4989 generic.go:334] "Generic (PLEG): container finished" podID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" containerID="7a2f0c49812f8cd37d298270e761a5be9be45b5bbd6517b8eaac1b2cff263220" exitCode=0 Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.435640 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xbw5q" event={"ID":"3bba6be3-7805-41e9-a2fb-d50fabe8b254","Type":"ContainerDied","Data":"7a2f0c49812f8cd37d298270e761a5be9be45b5bbd6517b8eaac1b2cff263220"} Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.436258 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.629732 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-catalog-content\") pod \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.629928 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txsmv\" (UniqueName: \"kubernetes.io/projected/f75af986-e7f9-48e9-8680-eb09e74cd9fb-kube-api-access-txsmv\") pod \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.629956 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-utilities\") pod \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\" (UID: \"f75af986-e7f9-48e9-8680-eb09e74cd9fb\") " Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.630766 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-utilities" (OuterVolumeSpecName: "utilities") pod "f75af986-e7f9-48e9-8680-eb09e74cd9fb" (UID: "f75af986-e7f9-48e9-8680-eb09e74cd9fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.636820 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f75af986-e7f9-48e9-8680-eb09e74cd9fb-kube-api-access-txsmv" (OuterVolumeSpecName: "kube-api-access-txsmv") pod "f75af986-e7f9-48e9-8680-eb09e74cd9fb" (UID: "f75af986-e7f9-48e9-8680-eb09e74cd9fb"). InnerVolumeSpecName "kube-api-access-txsmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.682181 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f75af986-e7f9-48e9-8680-eb09e74cd9fb" (UID: "f75af986-e7f9-48e9-8680-eb09e74cd9fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.745924 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txsmv\" (UniqueName: \"kubernetes.io/projected/f75af986-e7f9-48e9-8680-eb09e74cd9fb-kube-api-access-txsmv\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.745959 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:43 crc kubenswrapper[4989]: I1213 17:29:43.745977 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f75af986-e7f9-48e9-8680-eb09e74cd9fb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:44 crc kubenswrapper[4989]: I1213 17:29:44.446049 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xbw5q" event={"ID":"3bba6be3-7805-41e9-a2fb-d50fabe8b254","Type":"ContainerStarted","Data":"01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639"} Dec 13 17:29:44 crc kubenswrapper[4989]: I1213 17:29:44.446100 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqxjt" Dec 13 17:29:44 crc kubenswrapper[4989]: I1213 17:29:44.466629 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xbw5q" podStartSLOduration=2.935284514 podStartE2EDuration="5.466613344s" podCreationTimestamp="2025-12-13 17:29:39 +0000 UTC" firstStartedPulling="2025-12-13 17:29:41.422774293 +0000 UTC m=+796.029221431" lastFinishedPulling="2025-12-13 17:29:43.954103123 +0000 UTC m=+798.560550261" observedRunningTime="2025-12-13 17:29:44.464707434 +0000 UTC m=+799.071154612" watchObservedRunningTime="2025-12-13 17:29:44.466613344 +0000 UTC m=+799.073060482" Dec 13 17:29:44 crc kubenswrapper[4989]: I1213 17:29:44.489534 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hqxjt"] Dec 13 17:29:44 crc kubenswrapper[4989]: I1213 17:29:44.494601 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hqxjt"] Dec 13 17:29:45 crc kubenswrapper[4989]: I1213 17:29:45.219442 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:29:46 crc kubenswrapper[4989]: I1213 17:29:46.021508 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" path="/var/lib/kubelet/pods/f75af986-e7f9-48e9-8680-eb09e74cd9fb/volumes" Dec 13 17:29:48 crc kubenswrapper[4989]: I1213 17:29:48.918214 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cxz2g"] Dec 13 17:29:48 crc kubenswrapper[4989]: I1213 17:29:48.918636 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cxz2g" podUID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerName="registry-server" containerID="cri-o://96d6c2d24ec8588d0e62191e85b687e0b22d4d6cf3bc58b6c3632286f72c51fe" gracePeriod=2 Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.132753 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2pzwx"] Dec 13 17:29:49 crc kubenswrapper[4989]: E1213 17:29:49.133054 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" containerName="extract-utilities" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.133083 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" containerName="extract-utilities" Dec 13 17:29:49 crc kubenswrapper[4989]: E1213 17:29:49.133101 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" containerName="extract-content" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.133109 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" containerName="extract-content" Dec 13 17:29:49 crc kubenswrapper[4989]: E1213 17:29:49.133122 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" containerName="registry-server" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.133130 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" containerName="registry-server" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.133334 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f75af986-e7f9-48e9-8680-eb09e74cd9fb" containerName="registry-server" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.134376 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.145588 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2pzwx"] Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.214119 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2kpf\" (UniqueName: \"kubernetes.io/projected/45e513c6-d52f-4805-9e5c-ce787e388f8b-kube-api-access-l2kpf\") pod \"redhat-marketplace-2pzwx\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.214179 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-catalog-content\") pod \"redhat-marketplace-2pzwx\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.214209 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-utilities\") pod \"redhat-marketplace-2pzwx\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.315087 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-utilities\") pod \"redhat-marketplace-2pzwx\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.315190 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2kpf\" (UniqueName: \"kubernetes.io/projected/45e513c6-d52f-4805-9e5c-ce787e388f8b-kube-api-access-l2kpf\") pod \"redhat-marketplace-2pzwx\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.315220 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-catalog-content\") pod \"redhat-marketplace-2pzwx\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.315639 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-utilities\") pod \"redhat-marketplace-2pzwx\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.315703 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-catalog-content\") pod \"redhat-marketplace-2pzwx\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.359864 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2kpf\" (UniqueName: \"kubernetes.io/projected/45e513c6-d52f-4805-9e5c-ce787e388f8b-kube-api-access-l2kpf\") pod \"redhat-marketplace-2pzwx\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.485450 4989 generic.go:334] "Generic (PLEG): container finished" podID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerID="96d6c2d24ec8588d0e62191e85b687e0b22d4d6cf3bc58b6c3632286f72c51fe" exitCode=0 Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.485547 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxz2g" event={"ID":"f6c5fbf4-620a-45c5-85cd-04fb5f227247","Type":"ContainerDied","Data":"96d6c2d24ec8588d0e62191e85b687e0b22d4d6cf3bc58b6c3632286f72c51fe"} Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.492286 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.657822 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.658449 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.715479 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.792660 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.822638 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqrh2\" (UniqueName: \"kubernetes.io/projected/f6c5fbf4-620a-45c5-85cd-04fb5f227247-kube-api-access-vqrh2\") pod \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.822748 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-catalog-content\") pod \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.822806 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-utilities\") pod \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\" (UID: \"f6c5fbf4-620a-45c5-85cd-04fb5f227247\") " Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.824501 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-utilities" (OuterVolumeSpecName: "utilities") pod "f6c5fbf4-620a-45c5-85cd-04fb5f227247" (UID: "f6c5fbf4-620a-45c5-85cd-04fb5f227247"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.828657 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6c5fbf4-620a-45c5-85cd-04fb5f227247-kube-api-access-vqrh2" (OuterVolumeSpecName: "kube-api-access-vqrh2") pod "f6c5fbf4-620a-45c5-85cd-04fb5f227247" (UID: "f6c5fbf4-620a-45c5-85cd-04fb5f227247"). InnerVolumeSpecName "kube-api-access-vqrh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.923938 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqrh2\" (UniqueName: \"kubernetes.io/projected/f6c5fbf4-620a-45c5-85cd-04fb5f227247-kube-api-access-vqrh2\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.923967 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.943964 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6c5fbf4-620a-45c5-85cd-04fb5f227247" (UID: "f6c5fbf4-620a-45c5-85cd-04fb5f227247"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:29:49 crc kubenswrapper[4989]: I1213 17:29:49.959430 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2pzwx"] Dec 13 17:29:49 crc kubenswrapper[4989]: W1213 17:29:49.968157 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45e513c6_d52f_4805_9e5c_ce787e388f8b.slice/crio-a7394c17113dad8c038de15550208b387f54b8e4e52d222bbfbf4c3dfecb9e08 WatchSource:0}: Error finding container a7394c17113dad8c038de15550208b387f54b8e4e52d222bbfbf4c3dfecb9e08: Status 404 returned error can't find the container with id a7394c17113dad8c038de15550208b387f54b8e4e52d222bbfbf4c3dfecb9e08 Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.037592 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6c5fbf4-620a-45c5-85cd-04fb5f227247-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.493777 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxz2g" event={"ID":"f6c5fbf4-620a-45c5-85cd-04fb5f227247","Type":"ContainerDied","Data":"80f8c7e9356d9e8bd4dc108380120ed82417f19cb663c5dd3fd9b6818b86a531"} Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.493841 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxz2g" Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.493842 4989 scope.go:117] "RemoveContainer" containerID="96d6c2d24ec8588d0e62191e85b687e0b22d4d6cf3bc58b6c3632286f72c51fe" Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.496008 4989 generic.go:334] "Generic (PLEG): container finished" podID="45e513c6-d52f-4805-9e5c-ce787e388f8b" containerID="d9748ba92c2f9ffeb4b5a07a631e2675695c866137a710dda676cfda86345a03" exitCode=0 Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.496045 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2pzwx" event={"ID":"45e513c6-d52f-4805-9e5c-ce787e388f8b","Type":"ContainerDied","Data":"d9748ba92c2f9ffeb4b5a07a631e2675695c866137a710dda676cfda86345a03"} Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.496097 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2pzwx" event={"ID":"45e513c6-d52f-4805-9e5c-ce787e388f8b","Type":"ContainerStarted","Data":"a7394c17113dad8c038de15550208b387f54b8e4e52d222bbfbf4c3dfecb9e08"} Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.511712 4989 scope.go:117] "RemoveContainer" containerID="f5338fe03fe28709ebcc8d9503949d7a190f4e378b500c397caa9ac940a35146" Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.519481 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cxz2g"] Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.523331 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cxz2g"] Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.529605 4989 scope.go:117] "RemoveContainer" containerID="9c0c24711280c50a6caab7bb32276e7bbebb864a1f8561c7a1e31e8242f4701a" Dec 13 17:29:50 crc kubenswrapper[4989]: I1213 17:29:50.543156 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:51 crc kubenswrapper[4989]: I1213 17:29:51.504957 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2pzwx" event={"ID":"45e513c6-d52f-4805-9e5c-ce787e388f8b","Type":"ContainerStarted","Data":"c59f9bcaf1cd0ca0c015645adc631cfc4f890591c69a7ae5de3981d60e02acd3"} Dec 13 17:29:52 crc kubenswrapper[4989]: I1213 17:29:52.022096 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" path="/var/lib/kubelet/pods/f6c5fbf4-620a-45c5-85cd-04fb5f227247/volumes" Dec 13 17:29:52 crc kubenswrapper[4989]: I1213 17:29:52.514225 4989 generic.go:334] "Generic (PLEG): container finished" podID="45e513c6-d52f-4805-9e5c-ce787e388f8b" containerID="c59f9bcaf1cd0ca0c015645adc631cfc4f890591c69a7ae5de3981d60e02acd3" exitCode=0 Dec 13 17:29:52 crc kubenswrapper[4989]: I1213 17:29:52.514281 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2pzwx" event={"ID":"45e513c6-d52f-4805-9e5c-ce787e388f8b","Type":"ContainerDied","Data":"c59f9bcaf1cd0ca0c015645adc631cfc4f890591c69a7ae5de3981d60e02acd3"} Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.120597 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xbw5q"] Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.121098 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xbw5q" podUID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" containerName="registry-server" containerID="cri-o://01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639" gracePeriod=2 Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.505971 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.521895 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2pzwx" event={"ID":"45e513c6-d52f-4805-9e5c-ce787e388f8b","Type":"ContainerStarted","Data":"535f57e0f5c9b22f8ee7346072ded8f73751eb20fde366be0d788b3b2c8ca701"} Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.526326 4989 generic.go:334] "Generic (PLEG): container finished" podID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" containerID="01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639" exitCode=0 Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.526769 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xbw5q" event={"ID":"3bba6be3-7805-41e9-a2fb-d50fabe8b254","Type":"ContainerDied","Data":"01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639"} Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.526815 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xbw5q" event={"ID":"3bba6be3-7805-41e9-a2fb-d50fabe8b254","Type":"ContainerDied","Data":"e8a9dd9ad772b44197bbd8be8885dd2abc5f0ad005e7a57020ba67e16b3e7020"} Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.526836 4989 scope.go:117] "RemoveContainer" containerID="01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.526975 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xbw5q" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.543489 4989 scope.go:117] "RemoveContainer" containerID="7a2f0c49812f8cd37d298270e761a5be9be45b5bbd6517b8eaac1b2cff263220" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.545111 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2pzwx" podStartSLOduration=1.8122408970000001 podStartE2EDuration="4.545093085s" podCreationTimestamp="2025-12-13 17:29:49 +0000 UTC" firstStartedPulling="2025-12-13 17:29:50.497406911 +0000 UTC m=+805.103854049" lastFinishedPulling="2025-12-13 17:29:53.230259099 +0000 UTC m=+807.836706237" observedRunningTime="2025-12-13 17:29:53.543922558 +0000 UTC m=+808.150369706" watchObservedRunningTime="2025-12-13 17:29:53.545093085 +0000 UTC m=+808.151540223" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.562403 4989 scope.go:117] "RemoveContainer" containerID="56d4e750ad4a4e8bdd118ce322869a0fb4bafb1a369754fa4c515f341b90fb5a" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.577850 4989 scope.go:117] "RemoveContainer" containerID="01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639" Dec 13 17:29:53 crc kubenswrapper[4989]: E1213 17:29:53.578273 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639\": container with ID starting with 01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639 not found: ID does not exist" containerID="01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.578335 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639"} err="failed to get container status \"01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639\": rpc error: code = NotFound desc = could not find container \"01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639\": container with ID starting with 01e4223aca8f47fd1ab777afa7bd9efd5370675a79a52e56c761c846c0a1f639 not found: ID does not exist" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.578364 4989 scope.go:117] "RemoveContainer" containerID="7a2f0c49812f8cd37d298270e761a5be9be45b5bbd6517b8eaac1b2cff263220" Dec 13 17:29:53 crc kubenswrapper[4989]: E1213 17:29:53.579030 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a2f0c49812f8cd37d298270e761a5be9be45b5bbd6517b8eaac1b2cff263220\": container with ID starting with 7a2f0c49812f8cd37d298270e761a5be9be45b5bbd6517b8eaac1b2cff263220 not found: ID does not exist" containerID="7a2f0c49812f8cd37d298270e761a5be9be45b5bbd6517b8eaac1b2cff263220" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.579065 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a2f0c49812f8cd37d298270e761a5be9be45b5bbd6517b8eaac1b2cff263220"} err="failed to get container status \"7a2f0c49812f8cd37d298270e761a5be9be45b5bbd6517b8eaac1b2cff263220\": rpc error: code = NotFound desc = could not find container \"7a2f0c49812f8cd37d298270e761a5be9be45b5bbd6517b8eaac1b2cff263220\": container with ID starting with 7a2f0c49812f8cd37d298270e761a5be9be45b5bbd6517b8eaac1b2cff263220 not found: ID does not exist" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.579088 4989 scope.go:117] "RemoveContainer" containerID="56d4e750ad4a4e8bdd118ce322869a0fb4bafb1a369754fa4c515f341b90fb5a" Dec 13 17:29:53 crc kubenswrapper[4989]: E1213 17:29:53.579315 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56d4e750ad4a4e8bdd118ce322869a0fb4bafb1a369754fa4c515f341b90fb5a\": container with ID starting with 56d4e750ad4a4e8bdd118ce322869a0fb4bafb1a369754fa4c515f341b90fb5a not found: ID does not exist" containerID="56d4e750ad4a4e8bdd118ce322869a0fb4bafb1a369754fa4c515f341b90fb5a" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.579339 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56d4e750ad4a4e8bdd118ce322869a0fb4bafb1a369754fa4c515f341b90fb5a"} err="failed to get container status \"56d4e750ad4a4e8bdd118ce322869a0fb4bafb1a369754fa4c515f341b90fb5a\": rpc error: code = NotFound desc = could not find container \"56d4e750ad4a4e8bdd118ce322869a0fb4bafb1a369754fa4c515f341b90fb5a\": container with ID starting with 56d4e750ad4a4e8bdd118ce322869a0fb4bafb1a369754fa4c515f341b90fb5a not found: ID does not exist" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.591021 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-catalog-content\") pod \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.591065 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pwqp\" (UniqueName: \"kubernetes.io/projected/3bba6be3-7805-41e9-a2fb-d50fabe8b254-kube-api-access-8pwqp\") pod \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.591195 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-utilities\") pod \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\" (UID: \"3bba6be3-7805-41e9-a2fb-d50fabe8b254\") " Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.591889 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-utilities" (OuterVolumeSpecName: "utilities") pod "3bba6be3-7805-41e9-a2fb-d50fabe8b254" (UID: "3bba6be3-7805-41e9-a2fb-d50fabe8b254"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.596129 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bba6be3-7805-41e9-a2fb-d50fabe8b254-kube-api-access-8pwqp" (OuterVolumeSpecName: "kube-api-access-8pwqp") pod "3bba6be3-7805-41e9-a2fb-d50fabe8b254" (UID: "3bba6be3-7805-41e9-a2fb-d50fabe8b254"). InnerVolumeSpecName "kube-api-access-8pwqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.635481 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bba6be3-7805-41e9-a2fb-d50fabe8b254" (UID: "3bba6be3-7805-41e9-a2fb-d50fabe8b254"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.692880 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.692923 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pwqp\" (UniqueName: \"kubernetes.io/projected/3bba6be3-7805-41e9-a2fb-d50fabe8b254-kube-api-access-8pwqp\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.692938 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bba6be3-7805-41e9-a2fb-d50fabe8b254-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.854599 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xbw5q"] Dec 13 17:29:53 crc kubenswrapper[4989]: I1213 17:29:53.861966 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xbw5q"] Dec 13 17:29:54 crc kubenswrapper[4989]: I1213 17:29:54.021921 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" path="/var/lib/kubelet/pods/3bba6be3-7805-41e9-a2fb-d50fabe8b254/volumes" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.530092 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-48dg5"] Dec 13 17:29:56 crc kubenswrapper[4989]: E1213 17:29:56.530729 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerName="extract-content" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.530754 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerName="extract-content" Dec 13 17:29:56 crc kubenswrapper[4989]: E1213 17:29:56.530774 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerName="extract-utilities" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.530808 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerName="extract-utilities" Dec 13 17:29:56 crc kubenswrapper[4989]: E1213 17:29:56.530825 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" containerName="registry-server" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.530837 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" containerName="registry-server" Dec 13 17:29:56 crc kubenswrapper[4989]: E1213 17:29:56.530854 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" containerName="extract-content" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.530866 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" containerName="extract-content" Dec 13 17:29:56 crc kubenswrapper[4989]: E1213 17:29:56.530898 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerName="registry-server" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.530909 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerName="registry-server" Dec 13 17:29:56 crc kubenswrapper[4989]: E1213 17:29:56.530928 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" containerName="extract-utilities" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.530939 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" containerName="extract-utilities" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.531152 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bba6be3-7805-41e9-a2fb-d50fabe8b254" containerName="registry-server" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.531178 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6c5fbf4-620a-45c5-85cd-04fb5f227247" containerName="registry-server" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.531817 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-48dg5" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.534459 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-2m28j" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.539395 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-48dg5"] Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.632051 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfdz6\" (UniqueName: \"kubernetes.io/projected/b62f0d30-924c-4e82-9204-1e32c6f4008b-kube-api-access-qfdz6\") pod \"infra-operator-index-48dg5\" (UID: \"b62f0d30-924c-4e82-9204-1e32c6f4008b\") " pod="openstack-operators/infra-operator-index-48dg5" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.733420 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfdz6\" (UniqueName: \"kubernetes.io/projected/b62f0d30-924c-4e82-9204-1e32c6f4008b-kube-api-access-qfdz6\") pod \"infra-operator-index-48dg5\" (UID: \"b62f0d30-924c-4e82-9204-1e32c6f4008b\") " pod="openstack-operators/infra-operator-index-48dg5" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.753938 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfdz6\" (UniqueName: \"kubernetes.io/projected/b62f0d30-924c-4e82-9204-1e32c6f4008b-kube-api-access-qfdz6\") pod \"infra-operator-index-48dg5\" (UID: \"b62f0d30-924c-4e82-9204-1e32c6f4008b\") " pod="openstack-operators/infra-operator-index-48dg5" Dec 13 17:29:56 crc kubenswrapper[4989]: I1213 17:29:56.852285 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-48dg5" Dec 13 17:29:57 crc kubenswrapper[4989]: I1213 17:29:57.296557 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-48dg5"] Dec 13 17:29:57 crc kubenswrapper[4989]: W1213 17:29:57.317956 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb62f0d30_924c_4e82_9204_1e32c6f4008b.slice/crio-3fdce0456abc1bd6334ebe71e77ae54a76820557f4132a7bb15804e56852b1a9 WatchSource:0}: Error finding container 3fdce0456abc1bd6334ebe71e77ae54a76820557f4132a7bb15804e56852b1a9: Status 404 returned error can't find the container with id 3fdce0456abc1bd6334ebe71e77ae54a76820557f4132a7bb15804e56852b1a9 Dec 13 17:29:57 crc kubenswrapper[4989]: I1213 17:29:57.557865 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-48dg5" event={"ID":"b62f0d30-924c-4e82-9204-1e32c6f4008b","Type":"ContainerStarted","Data":"3fdce0456abc1bd6334ebe71e77ae54a76820557f4132a7bb15804e56852b1a9"} Dec 13 17:29:59 crc kubenswrapper[4989]: I1213 17:29:59.493536 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:59 crc kubenswrapper[4989]: I1213 17:29:59.493856 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:59 crc kubenswrapper[4989]: I1213 17:29:59.538199 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:29:59 crc kubenswrapper[4989]: I1213 17:29:59.620068 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.150773 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g"] Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.152609 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.155309 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.155361 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.158654 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g"] Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.220431 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85kv6\" (UniqueName: \"kubernetes.io/projected/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-kube-api-access-85kv6\") pod \"collect-profiles-29427450-5cs5g\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.220471 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-config-volume\") pod \"collect-profiles-29427450-5cs5g\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.220511 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-secret-volume\") pod \"collect-profiles-29427450-5cs5g\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.322031 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-config-volume\") pod \"collect-profiles-29427450-5cs5g\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.322099 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-secret-volume\") pod \"collect-profiles-29427450-5cs5g\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.322176 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85kv6\" (UniqueName: \"kubernetes.io/projected/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-kube-api-access-85kv6\") pod \"collect-profiles-29427450-5cs5g\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.323375 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-config-volume\") pod \"collect-profiles-29427450-5cs5g\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.327579 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-secret-volume\") pod \"collect-profiles-29427450-5cs5g\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.338036 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85kv6\" (UniqueName: \"kubernetes.io/projected/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-kube-api-access-85kv6\") pod \"collect-profiles-29427450-5cs5g\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:00 crc kubenswrapper[4989]: I1213 17:30:00.481355 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:01 crc kubenswrapper[4989]: I1213 17:30:01.922407 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2pzwx"] Dec 13 17:30:01 crc kubenswrapper[4989]: I1213 17:30:01.922665 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2pzwx" podUID="45e513c6-d52f-4805-9e5c-ce787e388f8b" containerName="registry-server" containerID="cri-o://535f57e0f5c9b22f8ee7346072ded8f73751eb20fde366be0d788b3b2c8ca701" gracePeriod=2 Dec 13 17:30:02 crc kubenswrapper[4989]: I1213 17:30:02.594126 4989 generic.go:334] "Generic (PLEG): container finished" podID="45e513c6-d52f-4805-9e5c-ce787e388f8b" containerID="535f57e0f5c9b22f8ee7346072ded8f73751eb20fde366be0d788b3b2c8ca701" exitCode=0 Dec 13 17:30:02 crc kubenswrapper[4989]: I1213 17:30:02.594166 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2pzwx" event={"ID":"45e513c6-d52f-4805-9e5c-ce787e388f8b","Type":"ContainerDied","Data":"535f57e0f5c9b22f8ee7346072ded8f73751eb20fde366be0d788b3b2c8ca701"} Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.557307 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.597468 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g"] Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.616812 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-catalog-content\") pod \"45e513c6-d52f-4805-9e5c-ce787e388f8b\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.617184 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2kpf\" (UniqueName: \"kubernetes.io/projected/45e513c6-d52f-4805-9e5c-ce787e388f8b-kube-api-access-l2kpf\") pod \"45e513c6-d52f-4805-9e5c-ce787e388f8b\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.617286 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-utilities\") pod \"45e513c6-d52f-4805-9e5c-ce787e388f8b\" (UID: \"45e513c6-d52f-4805-9e5c-ce787e388f8b\") " Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.618368 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-utilities" (OuterVolumeSpecName: "utilities") pod "45e513c6-d52f-4805-9e5c-ce787e388f8b" (UID: "45e513c6-d52f-4805-9e5c-ce787e388f8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.623692 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45e513c6-d52f-4805-9e5c-ce787e388f8b-kube-api-access-l2kpf" (OuterVolumeSpecName: "kube-api-access-l2kpf") pod "45e513c6-d52f-4805-9e5c-ce787e388f8b" (UID: "45e513c6-d52f-4805-9e5c-ce787e388f8b"). InnerVolumeSpecName "kube-api-access-l2kpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.628685 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" event={"ID":"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5","Type":"ContainerStarted","Data":"1233704371bd8ca1eaab43e25bfbce790e9e244de5e72331dad96c7be9201e1e"} Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.630771 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2pzwx" event={"ID":"45e513c6-d52f-4805-9e5c-ce787e388f8b","Type":"ContainerDied","Data":"a7394c17113dad8c038de15550208b387f54b8e4e52d222bbfbf4c3dfecb9e08"} Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.630841 4989 scope.go:117] "RemoveContainer" containerID="535f57e0f5c9b22f8ee7346072ded8f73751eb20fde366be0d788b3b2c8ca701" Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.630983 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2pzwx" Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.639196 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45e513c6-d52f-4805-9e5c-ce787e388f8b" (UID: "45e513c6-d52f-4805-9e5c-ce787e388f8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.646170 4989 scope.go:117] "RemoveContainer" containerID="c59f9bcaf1cd0ca0c015645adc631cfc4f890591c69a7ae5de3981d60e02acd3" Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.670706 4989 scope.go:117] "RemoveContainer" containerID="d9748ba92c2f9ffeb4b5a07a631e2675695c866137a710dda676cfda86345a03" Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.718348 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.718406 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2kpf\" (UniqueName: \"kubernetes.io/projected/45e513c6-d52f-4805-9e5c-ce787e388f8b-kube-api-access-l2kpf\") on node \"crc\" DevicePath \"\"" Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.718418 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45e513c6-d52f-4805-9e5c-ce787e388f8b-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.957421 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2pzwx"] Dec 13 17:30:05 crc kubenswrapper[4989]: I1213 17:30:05.962165 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2pzwx"] Dec 13 17:30:06 crc kubenswrapper[4989]: I1213 17:30:06.029449 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45e513c6-d52f-4805-9e5c-ce787e388f8b" path="/var/lib/kubelet/pods/45e513c6-d52f-4805-9e5c-ce787e388f8b/volumes" Dec 13 17:30:06 crc kubenswrapper[4989]: I1213 17:30:06.637551 4989 generic.go:334] "Generic (PLEG): container finished" podID="5ab6ab34-fa97-465e-b9c2-60a589a4d9c5" containerID="bb1fe1ce07f26abb00b5e6918d500f80143abc2315362e472a0ac6af1a5a0748" exitCode=0 Dec 13 17:30:06 crc kubenswrapper[4989]: I1213 17:30:06.637840 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" event={"ID":"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5","Type":"ContainerDied","Data":"bb1fe1ce07f26abb00b5e6918d500f80143abc2315362e472a0ac6af1a5a0748"} Dec 13 17:30:06 crc kubenswrapper[4989]: I1213 17:30:06.639377 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-48dg5" event={"ID":"b62f0d30-924c-4e82-9204-1e32c6f4008b","Type":"ContainerStarted","Data":"8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227"} Dec 13 17:30:06 crc kubenswrapper[4989]: I1213 17:30:06.677913 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-48dg5" podStartSLOduration=2.473797604 podStartE2EDuration="10.677896751s" podCreationTimestamp="2025-12-13 17:29:56 +0000 UTC" firstStartedPulling="2025-12-13 17:29:57.31958233 +0000 UTC m=+811.926029468" lastFinishedPulling="2025-12-13 17:30:05.523681477 +0000 UTC m=+820.130128615" observedRunningTime="2025-12-13 17:30:06.67500846 +0000 UTC m=+821.281455588" watchObservedRunningTime="2025-12-13 17:30:06.677896751 +0000 UTC m=+821.284343889" Dec 13 17:30:06 crc kubenswrapper[4989]: I1213 17:30:06.852854 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-48dg5" Dec 13 17:30:06 crc kubenswrapper[4989]: I1213 17:30:06.852895 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-48dg5" Dec 13 17:30:06 crc kubenswrapper[4989]: I1213 17:30:06.877642 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-48dg5" Dec 13 17:30:07 crc kubenswrapper[4989]: I1213 17:30:07.914095 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:07 crc kubenswrapper[4989]: I1213 17:30:07.942653 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-secret-volume\") pod \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " Dec 13 17:30:07 crc kubenswrapper[4989]: I1213 17:30:07.942758 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-config-volume\") pod \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " Dec 13 17:30:07 crc kubenswrapper[4989]: I1213 17:30:07.942853 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85kv6\" (UniqueName: \"kubernetes.io/projected/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-kube-api-access-85kv6\") pod \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\" (UID: \"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5\") " Dec 13 17:30:07 crc kubenswrapper[4989]: I1213 17:30:07.944188 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-config-volume" (OuterVolumeSpecName: "config-volume") pod "5ab6ab34-fa97-465e-b9c2-60a589a4d9c5" (UID: "5ab6ab34-fa97-465e-b9c2-60a589a4d9c5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:30:07 crc kubenswrapper[4989]: I1213 17:30:07.947241 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5ab6ab34-fa97-465e-b9c2-60a589a4d9c5" (UID: "5ab6ab34-fa97-465e-b9c2-60a589a4d9c5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:30:07 crc kubenswrapper[4989]: I1213 17:30:07.947894 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-kube-api-access-85kv6" (OuterVolumeSpecName: "kube-api-access-85kv6") pod "5ab6ab34-fa97-465e-b9c2-60a589a4d9c5" (UID: "5ab6ab34-fa97-465e-b9c2-60a589a4d9c5"). InnerVolumeSpecName "kube-api-access-85kv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:30:08 crc kubenswrapper[4989]: I1213 17:30:08.044317 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 17:30:08 crc kubenswrapper[4989]: I1213 17:30:08.044343 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85kv6\" (UniqueName: \"kubernetes.io/projected/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-kube-api-access-85kv6\") on node \"crc\" DevicePath \"\"" Dec 13 17:30:08 crc kubenswrapper[4989]: I1213 17:30:08.044358 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ab6ab34-fa97-465e-b9c2-60a589a4d9c5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 17:30:08 crc kubenswrapper[4989]: I1213 17:30:08.650963 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" Dec 13 17:30:08 crc kubenswrapper[4989]: I1213 17:30:08.650994 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29427450-5cs5g" event={"ID":"5ab6ab34-fa97-465e-b9c2-60a589a4d9c5","Type":"ContainerDied","Data":"1233704371bd8ca1eaab43e25bfbce790e9e244de5e72331dad96c7be9201e1e"} Dec 13 17:30:08 crc kubenswrapper[4989]: I1213 17:30:08.651074 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1233704371bd8ca1eaab43e25bfbce790e9e244de5e72331dad96c7be9201e1e" Dec 13 17:30:08 crc kubenswrapper[4989]: I1213 17:30:08.829485 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:30:08 crc kubenswrapper[4989]: I1213 17:30:08.829565 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:30:16 crc kubenswrapper[4989]: I1213 17:30:16.878776 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-48dg5" Dec 13 17:30:20 crc kubenswrapper[4989]: I1213 17:30:20.975602 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q"] Dec 13 17:30:20 crc kubenswrapper[4989]: E1213 17:30:20.976245 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e513c6-d52f-4805-9e5c-ce787e388f8b" containerName="extract-utilities" Dec 13 17:30:20 crc kubenswrapper[4989]: I1213 17:30:20.976266 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e513c6-d52f-4805-9e5c-ce787e388f8b" containerName="extract-utilities" Dec 13 17:30:20 crc kubenswrapper[4989]: E1213 17:30:20.976291 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab6ab34-fa97-465e-b9c2-60a589a4d9c5" containerName="collect-profiles" Dec 13 17:30:20 crc kubenswrapper[4989]: I1213 17:30:20.976303 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab6ab34-fa97-465e-b9c2-60a589a4d9c5" containerName="collect-profiles" Dec 13 17:30:20 crc kubenswrapper[4989]: E1213 17:30:20.976336 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e513c6-d52f-4805-9e5c-ce787e388f8b" containerName="registry-server" Dec 13 17:30:20 crc kubenswrapper[4989]: I1213 17:30:20.976349 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e513c6-d52f-4805-9e5c-ce787e388f8b" containerName="registry-server" Dec 13 17:30:20 crc kubenswrapper[4989]: E1213 17:30:20.976374 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e513c6-d52f-4805-9e5c-ce787e388f8b" containerName="extract-content" Dec 13 17:30:20 crc kubenswrapper[4989]: I1213 17:30:20.976386 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e513c6-d52f-4805-9e5c-ce787e388f8b" containerName="extract-content" Dec 13 17:30:20 crc kubenswrapper[4989]: I1213 17:30:20.976555 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="45e513c6-d52f-4805-9e5c-ce787e388f8b" containerName="registry-server" Dec 13 17:30:20 crc kubenswrapper[4989]: I1213 17:30:20.976580 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ab6ab34-fa97-465e-b9c2-60a589a4d9c5" containerName="collect-profiles" Dec 13 17:30:20 crc kubenswrapper[4989]: I1213 17:30:20.977909 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:20 crc kubenswrapper[4989]: I1213 17:30:20.980176 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k762t" Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:20.987494 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q"] Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.042066 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-util\") pod \"a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.042137 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hqfd\" (UniqueName: \"kubernetes.io/projected/492342cb-48b9-4cc7-ad59-1c4a815e665b-kube-api-access-4hqfd\") pod \"a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.042224 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-bundle\") pod \"a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.143489 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-bundle\") pod \"a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.143998 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-util\") pod \"a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.144048 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-bundle\") pod \"a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.144123 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hqfd\" (UniqueName: \"kubernetes.io/projected/492342cb-48b9-4cc7-ad59-1c4a815e665b-kube-api-access-4hqfd\") pod \"a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.144331 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-util\") pod \"a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.191926 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hqfd\" (UniqueName: \"kubernetes.io/projected/492342cb-48b9-4cc7-ad59-1c4a815e665b-kube-api-access-4hqfd\") pod \"a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.356910 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.540874 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q"] Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.725704 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" event={"ID":"492342cb-48b9-4cc7-ad59-1c4a815e665b","Type":"ContainerStarted","Data":"d6f7a9d8d20124ddb87de8aa31da75a829a05909e735e7459bed3a4ba493c25d"} Dec 13 17:30:21 crc kubenswrapper[4989]: I1213 17:30:21.725745 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" event={"ID":"492342cb-48b9-4cc7-ad59-1c4a815e665b","Type":"ContainerStarted","Data":"c2f09b49fd6acca84fee486fbce26e871dcb48d2dd3f3757bb9a48713d055ebe"} Dec 13 17:30:22 crc kubenswrapper[4989]: I1213 17:30:22.739618 4989 generic.go:334] "Generic (PLEG): container finished" podID="492342cb-48b9-4cc7-ad59-1c4a815e665b" containerID="d6f7a9d8d20124ddb87de8aa31da75a829a05909e735e7459bed3a4ba493c25d" exitCode=0 Dec 13 17:30:22 crc kubenswrapper[4989]: I1213 17:30:22.739661 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" event={"ID":"492342cb-48b9-4cc7-ad59-1c4a815e665b","Type":"ContainerDied","Data":"d6f7a9d8d20124ddb87de8aa31da75a829a05909e735e7459bed3a4ba493c25d"} Dec 13 17:30:24 crc kubenswrapper[4989]: I1213 17:30:24.755774 4989 generic.go:334] "Generic (PLEG): container finished" podID="492342cb-48b9-4cc7-ad59-1c4a815e665b" containerID="74a78c7c620e342a5341cdc87224b5e81a96ccf2a68d1f5cd518274d58301291" exitCode=0 Dec 13 17:30:24 crc kubenswrapper[4989]: I1213 17:30:24.755900 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" event={"ID":"492342cb-48b9-4cc7-ad59-1c4a815e665b","Type":"ContainerDied","Data":"74a78c7c620e342a5341cdc87224b5e81a96ccf2a68d1f5cd518274d58301291"} Dec 13 17:30:25 crc kubenswrapper[4989]: I1213 17:30:25.768120 4989 generic.go:334] "Generic (PLEG): container finished" podID="492342cb-48b9-4cc7-ad59-1c4a815e665b" containerID="f5d55af33e4e0fa93a533a9ae15179cf09fb9b6818ef13896df5a8e0ee51230a" exitCode=0 Dec 13 17:30:25 crc kubenswrapper[4989]: I1213 17:30:25.768246 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" event={"ID":"492342cb-48b9-4cc7-ad59-1c4a815e665b","Type":"ContainerDied","Data":"f5d55af33e4e0fa93a533a9ae15179cf09fb9b6818ef13896df5a8e0ee51230a"} Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.073339 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.234106 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-bundle\") pod \"492342cb-48b9-4cc7-ad59-1c4a815e665b\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.234181 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hqfd\" (UniqueName: \"kubernetes.io/projected/492342cb-48b9-4cc7-ad59-1c4a815e665b-kube-api-access-4hqfd\") pod \"492342cb-48b9-4cc7-ad59-1c4a815e665b\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.234215 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-util\") pod \"492342cb-48b9-4cc7-ad59-1c4a815e665b\" (UID: \"492342cb-48b9-4cc7-ad59-1c4a815e665b\") " Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.236486 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-bundle" (OuterVolumeSpecName: "bundle") pod "492342cb-48b9-4cc7-ad59-1c4a815e665b" (UID: "492342cb-48b9-4cc7-ad59-1c4a815e665b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.245150 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/492342cb-48b9-4cc7-ad59-1c4a815e665b-kube-api-access-4hqfd" (OuterVolumeSpecName: "kube-api-access-4hqfd") pod "492342cb-48b9-4cc7-ad59-1c4a815e665b" (UID: "492342cb-48b9-4cc7-ad59-1c4a815e665b"). InnerVolumeSpecName "kube-api-access-4hqfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.283418 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-util" (OuterVolumeSpecName: "util") pod "492342cb-48b9-4cc7-ad59-1c4a815e665b" (UID: "492342cb-48b9-4cc7-ad59-1c4a815e665b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.335678 4989 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.335709 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hqfd\" (UniqueName: \"kubernetes.io/projected/492342cb-48b9-4cc7-ad59-1c4a815e665b-kube-api-access-4hqfd\") on node \"crc\" DevicePath \"\"" Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.335723 4989 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/492342cb-48b9-4cc7-ad59-1c4a815e665b-util\") on node \"crc\" DevicePath \"\"" Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.785535 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" event={"ID":"492342cb-48b9-4cc7-ad59-1c4a815e665b","Type":"ContainerDied","Data":"c2f09b49fd6acca84fee486fbce26e871dcb48d2dd3f3757bb9a48713d055ebe"} Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.785579 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2f09b49fd6acca84fee486fbce26e871dcb48d2dd3f3757bb9a48713d055ebe" Dec 13 17:30:27 crc kubenswrapper[4989]: I1213 17:30:27.785638 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.275290 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6"] Dec 13 17:30:38 crc kubenswrapper[4989]: E1213 17:30:38.276486 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="492342cb-48b9-4cc7-ad59-1c4a815e665b" containerName="extract" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.276506 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="492342cb-48b9-4cc7-ad59-1c4a815e665b" containerName="extract" Dec 13 17:30:38 crc kubenswrapper[4989]: E1213 17:30:38.276541 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="492342cb-48b9-4cc7-ad59-1c4a815e665b" containerName="pull" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.276549 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="492342cb-48b9-4cc7-ad59-1c4a815e665b" containerName="pull" Dec 13 17:30:38 crc kubenswrapper[4989]: E1213 17:30:38.276560 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="492342cb-48b9-4cc7-ad59-1c4a815e665b" containerName="util" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.276570 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="492342cb-48b9-4cc7-ad59-1c4a815e665b" containerName="util" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.276710 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="492342cb-48b9-4cc7-ad59-1c4a815e665b" containerName="extract" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.277410 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.280501 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-z6mwn" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.280548 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.293439 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6"] Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.400907 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-apiservice-cert\") pod \"infra-operator-controller-manager-6db5fd65d9-ct8s6\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.400965 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrcdb\" (UniqueName: \"kubernetes.io/projected/fbe7e3ff-0c94-4738-9dfc-577938b357d2-kube-api-access-jrcdb\") pod \"infra-operator-controller-manager-6db5fd65d9-ct8s6\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.400994 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-webhook-cert\") pod \"infra-operator-controller-manager-6db5fd65d9-ct8s6\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.502139 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrcdb\" (UniqueName: \"kubernetes.io/projected/fbe7e3ff-0c94-4738-9dfc-577938b357d2-kube-api-access-jrcdb\") pod \"infra-operator-controller-manager-6db5fd65d9-ct8s6\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.502182 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-webhook-cert\") pod \"infra-operator-controller-manager-6db5fd65d9-ct8s6\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.502248 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-apiservice-cert\") pod \"infra-operator-controller-manager-6db5fd65d9-ct8s6\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.514675 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-webhook-cert\") pod \"infra-operator-controller-manager-6db5fd65d9-ct8s6\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.514936 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-apiservice-cert\") pod \"infra-operator-controller-manager-6db5fd65d9-ct8s6\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.521554 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrcdb\" (UniqueName: \"kubernetes.io/projected/fbe7e3ff-0c94-4738-9dfc-577938b357d2-kube-api-access-jrcdb\") pod \"infra-operator-controller-manager-6db5fd65d9-ct8s6\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.597503 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.829862 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.830098 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.830147 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.830734 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b6113f2a3ccfb4b50af5284bdcbd31c249a42a87c9815c957aa5451a83308bc9"} pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 17:30:38 crc kubenswrapper[4989]: I1213 17:30:38.830811 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" containerID="cri-o://b6113f2a3ccfb4b50af5284bdcbd31c249a42a87c9815c957aa5451a83308bc9" gracePeriod=600 Dec 13 17:30:39 crc kubenswrapper[4989]: I1213 17:30:39.040192 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6"] Dec 13 17:30:39 crc kubenswrapper[4989]: I1213 17:30:39.857817 4989 generic.go:334] "Generic (PLEG): container finished" podID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerID="b6113f2a3ccfb4b50af5284bdcbd31c249a42a87c9815c957aa5451a83308bc9" exitCode=0 Dec 13 17:30:39 crc kubenswrapper[4989]: I1213 17:30:39.857865 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerDied","Data":"b6113f2a3ccfb4b50af5284bdcbd31c249a42a87c9815c957aa5451a83308bc9"} Dec 13 17:30:39 crc kubenswrapper[4989]: I1213 17:30:39.858143 4989 scope.go:117] "RemoveContainer" containerID="c77b5f0a81370cabedf086384605132d2ee7a5c7c37053b21e73aede752d2782" Dec 13 17:30:39 crc kubenswrapper[4989]: I1213 17:30:39.859432 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" event={"ID":"fbe7e3ff-0c94-4738-9dfc-577938b357d2","Type":"ContainerStarted","Data":"d97a167ced38ed806aa2cb9b0f4f8369d635b3aaf8d67dd78c13ff1908a74696"} Dec 13 17:30:41 crc kubenswrapper[4989]: I1213 17:30:41.871650 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerStarted","Data":"f7a34899e26b30f2d8d01fb7d7814d99ead88a61514d27cbbc8a2e72806061a9"} Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.539664 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.541202 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.543036 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openshift-service-ca.crt" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.544445 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openstack-config-data" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.544502 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"kube-root-ca.crt" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.545055 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"galera-openstack-dockercfg-gvwv4" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.545053 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openstack-scripts" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.545505 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.546629 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.557148 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.558213 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.563588 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.569264 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.580277 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.665625 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kolla-config\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.665670 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.665690 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-kolla-config\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.665708 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-default\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.665728 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.665931 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-default\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.665987 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95t6f\" (UniqueName: \"kubernetes.io/projected/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kube-api-access-95t6f\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.666046 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-default\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.666080 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-generated\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.666130 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7bww\" (UniqueName: \"kubernetes.io/projected/b50061aa-30a3-4145-bca8-94c802594dc4-kube-api-access-l7bww\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.666221 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.666262 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-generated\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.666346 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-operator-scripts\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.666388 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7chb\" (UniqueName: \"kubernetes.io/projected/91eeba9c-4026-408d-9547-3bb58ae605dd-kube-api-access-q7chb\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.666428 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-operator-scripts\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.666471 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.666507 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-kolla-config\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.666528 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.767705 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-kolla-config\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.767765 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-default\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.767807 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.767884 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-default\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.767909 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95t6f\" (UniqueName: \"kubernetes.io/projected/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kube-api-access-95t6f\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.767940 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-default\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.767991 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-generated\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768023 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7bww\" (UniqueName: \"kubernetes.io/projected/b50061aa-30a3-4145-bca8-94c802594dc4-kube-api-access-l7bww\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768074 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768095 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-generated\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768152 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-operator-scripts\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768182 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7chb\" (UniqueName: \"kubernetes.io/projected/91eeba9c-4026-408d-9547-3bb58ae605dd-kube-api-access-q7chb\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768203 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-operator-scripts\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768227 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768228 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") device mount path \"/mnt/openstack/pv03\"" pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768249 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-kolla-config\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768480 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768563 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kolla-config\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768584 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.768778 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") device mount path \"/mnt/openstack/pv04\"" pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.769746 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-generated\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.770084 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-default\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.770274 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-default\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.770500 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-kolla-config\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.770629 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") device mount path \"/mnt/openstack/pv01\"" pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.771239 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-operator-scripts\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.783465 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kolla-config\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.784743 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-kolla-config\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.785880 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-default\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.786250 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-generated\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.787167 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-generated\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.803204 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7chb\" (UniqueName: \"kubernetes.io/projected/91eeba9c-4026-408d-9547-3bb58ae605dd-kube-api-access-q7chb\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.803364 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95t6f\" (UniqueName: \"kubernetes.io/projected/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kube-api-access-95t6f\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.804947 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-operator-scripts\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.806007 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-operator-scripts\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.812714 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7bww\" (UniqueName: \"kubernetes.io/projected/b50061aa-30a3-4145-bca8-94c802594dc4-kube-api-access-l7bww\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.817257 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.819899 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.820098 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.869394 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.886002 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:30:42 crc kubenswrapper[4989]: I1213 17:30:42.896767 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:30:43 crc kubenswrapper[4989]: I1213 17:30:43.603956 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 13 17:30:43 crc kubenswrapper[4989]: I1213 17:30:43.607912 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 13 17:30:43 crc kubenswrapper[4989]: W1213 17:30:43.613111 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc540cd7f_4b50_4a71_9e51_0687156ee1b2.slice/crio-1c866768a930d535c9685f7938676ec6e51e35edb7ff02f9c5bf1d628f6b75ff WatchSource:0}: Error finding container 1c866768a930d535c9685f7938676ec6e51e35edb7ff02f9c5bf1d628f6b75ff: Status 404 returned error can't find the container with id 1c866768a930d535c9685f7938676ec6e51e35edb7ff02f9c5bf1d628f6b75ff Dec 13 17:30:43 crc kubenswrapper[4989]: I1213 17:30:43.625098 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 13 17:30:43 crc kubenswrapper[4989]: W1213 17:30:43.626657 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91eeba9c_4026_408d_9547_3bb58ae605dd.slice/crio-34d5cc57105ffa0cae7f18c0be8a48edd2c7bb7f927e152406a6f5cf911fd741 WatchSource:0}: Error finding container 34d5cc57105ffa0cae7f18c0be8a48edd2c7bb7f927e152406a6f5cf911fd741: Status 404 returned error can't find the container with id 34d5cc57105ffa0cae7f18c0be8a48edd2c7bb7f927e152406a6f5cf911fd741 Dec 13 17:30:43 crc kubenswrapper[4989]: I1213 17:30:43.884678 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"b50061aa-30a3-4145-bca8-94c802594dc4","Type":"ContainerStarted","Data":"d89dc977938d99ccded63ad6416b99bb492b32eda5f4ea10635f6d3f862cd994"} Dec 13 17:30:43 crc kubenswrapper[4989]: I1213 17:30:43.885514 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"c540cd7f-4b50-4a71-9e51-0687156ee1b2","Type":"ContainerStarted","Data":"1c866768a930d535c9685f7938676ec6e51e35edb7ff02f9c5bf1d628f6b75ff"} Dec 13 17:30:43 crc kubenswrapper[4989]: I1213 17:30:43.886174 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"91eeba9c-4026-408d-9547-3bb58ae605dd","Type":"ContainerStarted","Data":"34d5cc57105ffa0cae7f18c0be8a48edd2c7bb7f927e152406a6f5cf911fd741"} Dec 13 17:30:43 crc kubenswrapper[4989]: I1213 17:30:43.887243 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" event={"ID":"fbe7e3ff-0c94-4738-9dfc-577938b357d2","Type":"ContainerStarted","Data":"505d75199ebd00bcb7d5e147020cb3c303472d6032b5f7e84d0071a795e7a877"} Dec 13 17:30:43 crc kubenswrapper[4989]: I1213 17:30:43.888067 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:43 crc kubenswrapper[4989]: I1213 17:30:43.912454 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" podStartSLOduration=1.784425476 podStartE2EDuration="5.912436107s" podCreationTimestamp="2025-12-13 17:30:38 +0000 UTC" firstStartedPulling="2025-12-13 17:30:39.046322069 +0000 UTC m=+853.652769227" lastFinishedPulling="2025-12-13 17:30:43.17433272 +0000 UTC m=+857.780779858" observedRunningTime="2025-12-13 17:30:43.908286759 +0000 UTC m=+858.514733897" watchObservedRunningTime="2025-12-13 17:30:43.912436107 +0000 UTC m=+858.518883245" Dec 13 17:30:48 crc kubenswrapper[4989]: I1213 17:30:48.602375 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:30:51 crc kubenswrapper[4989]: I1213 17:30:51.964264 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"b50061aa-30a3-4145-bca8-94c802594dc4","Type":"ContainerStarted","Data":"34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec"} Dec 13 17:30:51 crc kubenswrapper[4989]: I1213 17:30:51.966571 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"c540cd7f-4b50-4a71-9e51-0687156ee1b2","Type":"ContainerStarted","Data":"df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6"} Dec 13 17:30:51 crc kubenswrapper[4989]: I1213 17:30:51.967809 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"91eeba9c-4026-408d-9547-3bb58ae605dd","Type":"ContainerStarted","Data":"9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517"} Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.190906 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.192896 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.194776 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"memcached-config-data" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.195282 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"memcached-memcached-dockercfg-pbtcl" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.210689 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.244224 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rml4\" (UniqueName: \"kubernetes.io/projected/bb43b605-721a-4dae-98c8-eeabd32ba579-kube-api-access-8rml4\") pod \"memcached-0\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " pod="cinder-kuttl-tests/memcached-0" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.244273 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-config-data\") pod \"memcached-0\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " pod="cinder-kuttl-tests/memcached-0" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.244344 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-kolla-config\") pod \"memcached-0\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " pod="cinder-kuttl-tests/memcached-0" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.345927 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-kolla-config\") pod \"memcached-0\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " pod="cinder-kuttl-tests/memcached-0" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.346022 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rml4\" (UniqueName: \"kubernetes.io/projected/bb43b605-721a-4dae-98c8-eeabd32ba579-kube-api-access-8rml4\") pod \"memcached-0\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " pod="cinder-kuttl-tests/memcached-0" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.346042 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-config-data\") pod \"memcached-0\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " pod="cinder-kuttl-tests/memcached-0" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.346916 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-kolla-config\") pod \"memcached-0\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " pod="cinder-kuttl-tests/memcached-0" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.347022 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-config-data\") pod \"memcached-0\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " pod="cinder-kuttl-tests/memcached-0" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.364603 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rml4\" (UniqueName: \"kubernetes.io/projected/bb43b605-721a-4dae-98c8-eeabd32ba579-kube-api-access-8rml4\") pod \"memcached-0\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " pod="cinder-kuttl-tests/memcached-0" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.513441 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.723164 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-wtqpt"] Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.724101 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.726010 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-4n6rs" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.817255 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-wtqpt"] Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.855519 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q7mb\" (UniqueName: \"kubernetes.io/projected/0fa9d119-2c72-4737-9eb7-8de2b492f714-kube-api-access-5q7mb\") pod \"rabbitmq-cluster-operator-index-wtqpt\" (UID: \"0fa9d119-2c72-4737-9eb7-8de2b492f714\") " pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.956913 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q7mb\" (UniqueName: \"kubernetes.io/projected/0fa9d119-2c72-4737-9eb7-8de2b492f714-kube-api-access-5q7mb\") pod \"rabbitmq-cluster-operator-index-wtqpt\" (UID: \"0fa9d119-2c72-4737-9eb7-8de2b492f714\") " pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.968284 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.984476 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q7mb\" (UniqueName: \"kubernetes.io/projected/0fa9d119-2c72-4737-9eb7-8de2b492f714-kube-api-access-5q7mb\") pod \"rabbitmq-cluster-operator-index-wtqpt\" (UID: \"0fa9d119-2c72-4737-9eb7-8de2b492f714\") " pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" Dec 13 17:30:54 crc kubenswrapper[4989]: I1213 17:30:54.998543 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"bb43b605-721a-4dae-98c8-eeabd32ba579","Type":"ContainerStarted","Data":"cce1c659efffb47c75f85a472e8c0325c4434956e81648b29411f03894bce5a3"} Dec 13 17:30:55 crc kubenswrapper[4989]: I1213 17:30:55.058201 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" Dec 13 17:30:55 crc kubenswrapper[4989]: I1213 17:30:55.370504 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-wtqpt"] Dec 13 17:30:55 crc kubenswrapper[4989]: W1213 17:30:55.386516 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fa9d119_2c72_4737_9eb7_8de2b492f714.slice/crio-3a2680737ea99929e54b64fd00668b3d2f368fee98375438086bc6d56e3e4f9b WatchSource:0}: Error finding container 3a2680737ea99929e54b64fd00668b3d2f368fee98375438086bc6d56e3e4f9b: Status 404 returned error can't find the container with id 3a2680737ea99929e54b64fd00668b3d2f368fee98375438086bc6d56e3e4f9b Dec 13 17:30:56 crc kubenswrapper[4989]: I1213 17:30:56.023654 4989 generic.go:334] "Generic (PLEG): container finished" podID="b50061aa-30a3-4145-bca8-94c802594dc4" containerID="34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec" exitCode=0 Dec 13 17:30:56 crc kubenswrapper[4989]: I1213 17:30:56.025638 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"b50061aa-30a3-4145-bca8-94c802594dc4","Type":"ContainerDied","Data":"34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec"} Dec 13 17:30:56 crc kubenswrapper[4989]: I1213 17:30:56.025686 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" event={"ID":"0fa9d119-2c72-4737-9eb7-8de2b492f714","Type":"ContainerStarted","Data":"3a2680737ea99929e54b64fd00668b3d2f368fee98375438086bc6d56e3e4f9b"} Dec 13 17:30:56 crc kubenswrapper[4989]: I1213 17:30:56.028644 4989 generic.go:334] "Generic (PLEG): container finished" podID="c540cd7f-4b50-4a71-9e51-0687156ee1b2" containerID="df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6" exitCode=0 Dec 13 17:30:56 crc kubenswrapper[4989]: I1213 17:30:56.028712 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"c540cd7f-4b50-4a71-9e51-0687156ee1b2","Type":"ContainerDied","Data":"df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6"} Dec 13 17:30:56 crc kubenswrapper[4989]: I1213 17:30:56.031021 4989 generic.go:334] "Generic (PLEG): container finished" podID="91eeba9c-4026-408d-9547-3bb58ae605dd" containerID="9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517" exitCode=0 Dec 13 17:30:56 crc kubenswrapper[4989]: I1213 17:30:56.031053 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"91eeba9c-4026-408d-9547-3bb58ae605dd","Type":"ContainerDied","Data":"9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517"} Dec 13 17:30:57 crc kubenswrapper[4989]: I1213 17:30:57.037646 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"91eeba9c-4026-408d-9547-3bb58ae605dd","Type":"ContainerStarted","Data":"b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4"} Dec 13 17:30:57 crc kubenswrapper[4989]: I1213 17:30:57.048006 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"b50061aa-30a3-4145-bca8-94c802594dc4","Type":"ContainerStarted","Data":"d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c"} Dec 13 17:30:57 crc kubenswrapper[4989]: I1213 17:30:57.053264 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"c540cd7f-4b50-4a71-9e51-0687156ee1b2","Type":"ContainerStarted","Data":"28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c"} Dec 13 17:30:57 crc kubenswrapper[4989]: I1213 17:30:57.082391 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-0" podStartSLOduration=8.054908445 podStartE2EDuration="16.082373621s" podCreationTimestamp="2025-12-13 17:30:41 +0000 UTC" firstStartedPulling="2025-12-13 17:30:43.628866694 +0000 UTC m=+858.235313822" lastFinishedPulling="2025-12-13 17:30:51.65633186 +0000 UTC m=+866.262778998" observedRunningTime="2025-12-13 17:30:57.060534498 +0000 UTC m=+871.666981656" watchObservedRunningTime="2025-12-13 17:30:57.082373621 +0000 UTC m=+871.688820779" Dec 13 17:30:57 crc kubenswrapper[4989]: I1213 17:30:57.085068 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-1" podStartSLOduration=8.018070769 podStartE2EDuration="16.085051644s" podCreationTimestamp="2025-12-13 17:30:41 +0000 UTC" firstStartedPulling="2025-12-13 17:30:43.613627384 +0000 UTC m=+858.220074522" lastFinishedPulling="2025-12-13 17:30:51.680608259 +0000 UTC m=+866.287055397" observedRunningTime="2025-12-13 17:30:57.078559104 +0000 UTC m=+871.685006262" watchObservedRunningTime="2025-12-13 17:30:57.085051644 +0000 UTC m=+871.691498792" Dec 13 17:30:57 crc kubenswrapper[4989]: I1213 17:30:57.099626 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-2" podStartSLOduration=8.079304427 podStartE2EDuration="16.099605753s" podCreationTimestamp="2025-12-13 17:30:41 +0000 UTC" firstStartedPulling="2025-12-13 17:30:43.616003697 +0000 UTC m=+858.222450825" lastFinishedPulling="2025-12-13 17:30:51.636305013 +0000 UTC m=+866.242752151" observedRunningTime="2025-12-13 17:30:57.093947628 +0000 UTC m=+871.700394766" watchObservedRunningTime="2025-12-13 17:30:57.099605753 +0000 UTC m=+871.706052901" Dec 13 17:30:58 crc kubenswrapper[4989]: I1213 17:30:58.908712 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-wtqpt"] Dec 13 17:30:59 crc kubenswrapper[4989]: I1213 17:30:59.516248 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-59pgz"] Dec 13 17:30:59 crc kubenswrapper[4989]: I1213 17:30:59.517308 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" Dec 13 17:30:59 crc kubenswrapper[4989]: I1213 17:30:59.525537 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-59pgz"] Dec 13 17:30:59 crc kubenswrapper[4989]: I1213 17:30:59.629865 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knfzn\" (UniqueName: \"kubernetes.io/projected/f0c708a3-be93-496d-85d5-fdedb653aa0d-kube-api-access-knfzn\") pod \"rabbitmq-cluster-operator-index-59pgz\" (UID: \"f0c708a3-be93-496d-85d5-fdedb653aa0d\") " pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" Dec 13 17:30:59 crc kubenswrapper[4989]: I1213 17:30:59.731571 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knfzn\" (UniqueName: \"kubernetes.io/projected/f0c708a3-be93-496d-85d5-fdedb653aa0d-kube-api-access-knfzn\") pod \"rabbitmq-cluster-operator-index-59pgz\" (UID: \"f0c708a3-be93-496d-85d5-fdedb653aa0d\") " pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" Dec 13 17:30:59 crc kubenswrapper[4989]: I1213 17:30:59.755660 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knfzn\" (UniqueName: \"kubernetes.io/projected/f0c708a3-be93-496d-85d5-fdedb653aa0d-kube-api-access-knfzn\") pod \"rabbitmq-cluster-operator-index-59pgz\" (UID: \"f0c708a3-be93-496d-85d5-fdedb653aa0d\") " pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" Dec 13 17:30:59 crc kubenswrapper[4989]: I1213 17:30:59.844392 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" Dec 13 17:31:02 crc kubenswrapper[4989]: I1213 17:31:02.872920 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:31:02 crc kubenswrapper[4989]: I1213 17:31:02.874832 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:31:02 crc kubenswrapper[4989]: I1213 17:31:02.888421 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:31:02 crc kubenswrapper[4989]: I1213 17:31:02.888461 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:31:02 crc kubenswrapper[4989]: I1213 17:31:02.899363 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:31:02 crc kubenswrapper[4989]: I1213 17:31:02.899427 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:31:03 crc kubenswrapper[4989]: I1213 17:31:03.123550 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/memcached-0" Dec 13 17:31:03 crc kubenswrapper[4989]: I1213 17:31:03.143901 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/memcached-0" podStartSLOduration=1.364874959 podStartE2EDuration="9.143880725s" podCreationTimestamp="2025-12-13 17:30:54 +0000 UTC" firstStartedPulling="2025-12-13 17:30:54.983462979 +0000 UTC m=+869.589910117" lastFinishedPulling="2025-12-13 17:31:02.762468725 +0000 UTC m=+877.368915883" observedRunningTime="2025-12-13 17:31:03.142365089 +0000 UTC m=+877.748812227" watchObservedRunningTime="2025-12-13 17:31:03.143880725 +0000 UTC m=+877.750327873" Dec 13 17:31:03 crc kubenswrapper[4989]: I1213 17:31:03.307249 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-59pgz"] Dec 13 17:31:04 crc kubenswrapper[4989]: I1213 17:31:04.130449 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"bb43b605-721a-4dae-98c8-eeabd32ba579","Type":"ContainerStarted","Data":"53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf"} Dec 13 17:31:04 crc kubenswrapper[4989]: I1213 17:31:04.132311 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" event={"ID":"0fa9d119-2c72-4737-9eb7-8de2b492f714","Type":"ContainerStarted","Data":"7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88"} Dec 13 17:31:04 crc kubenswrapper[4989]: I1213 17:31:04.132421 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" podUID="0fa9d119-2c72-4737-9eb7-8de2b492f714" containerName="registry-server" containerID="cri-o://7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88" gracePeriod=2 Dec 13 17:31:04 crc kubenswrapper[4989]: I1213 17:31:04.135348 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" event={"ID":"f0c708a3-be93-496d-85d5-fdedb653aa0d","Type":"ContainerStarted","Data":"cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff"} Dec 13 17:31:04 crc kubenswrapper[4989]: I1213 17:31:04.135378 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" event={"ID":"f0c708a3-be93-496d-85d5-fdedb653aa0d","Type":"ContainerStarted","Data":"24617ac4319df9250811f04c6b64b0ef0a5132ae7477f785ebc8326f64c15b8b"} Dec 13 17:31:04 crc kubenswrapper[4989]: I1213 17:31:04.168939 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" podStartSLOduration=2.67849543 podStartE2EDuration="10.168913378s" podCreationTimestamp="2025-12-13 17:30:54 +0000 UTC" firstStartedPulling="2025-12-13 17:30:55.389585511 +0000 UTC m=+869.996032649" lastFinishedPulling="2025-12-13 17:31:02.880003459 +0000 UTC m=+877.486450597" observedRunningTime="2025-12-13 17:31:04.16410318 +0000 UTC m=+878.770550318" watchObservedRunningTime="2025-12-13 17:31:04.168913378 +0000 UTC m=+878.775360526" Dec 13 17:31:04 crc kubenswrapper[4989]: I1213 17:31:04.186130 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" podStartSLOduration=4.732649997 podStartE2EDuration="5.186097778s" podCreationTimestamp="2025-12-13 17:30:59 +0000 UTC" firstStartedPulling="2025-12-13 17:31:03.325400281 +0000 UTC m=+877.931847419" lastFinishedPulling="2025-12-13 17:31:03.778848062 +0000 UTC m=+878.385295200" observedRunningTime="2025-12-13 17:31:04.182056203 +0000 UTC m=+878.788503351" watchObservedRunningTime="2025-12-13 17:31:04.186097778 +0000 UTC m=+878.792544916" Dec 13 17:31:04 crc kubenswrapper[4989]: I1213 17:31:04.518704 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" Dec 13 17:31:04 crc kubenswrapper[4989]: I1213 17:31:04.611761 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q7mb\" (UniqueName: \"kubernetes.io/projected/0fa9d119-2c72-4737-9eb7-8de2b492f714-kube-api-access-5q7mb\") pod \"0fa9d119-2c72-4737-9eb7-8de2b492f714\" (UID: \"0fa9d119-2c72-4737-9eb7-8de2b492f714\") " Dec 13 17:31:04 crc kubenswrapper[4989]: I1213 17:31:04.618288 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fa9d119-2c72-4737-9eb7-8de2b492f714-kube-api-access-5q7mb" (OuterVolumeSpecName: "kube-api-access-5q7mb") pod "0fa9d119-2c72-4737-9eb7-8de2b492f714" (UID: "0fa9d119-2c72-4737-9eb7-8de2b492f714"). InnerVolumeSpecName "kube-api-access-5q7mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:31:04 crc kubenswrapper[4989]: I1213 17:31:04.713964 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q7mb\" (UniqueName: \"kubernetes.io/projected/0fa9d119-2c72-4737-9eb7-8de2b492f714-kube-api-access-5q7mb\") on node \"crc\" DevicePath \"\"" Dec 13 17:31:05 crc kubenswrapper[4989]: I1213 17:31:05.143406 4989 generic.go:334] "Generic (PLEG): container finished" podID="0fa9d119-2c72-4737-9eb7-8de2b492f714" containerID="7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88" exitCode=0 Dec 13 17:31:05 crc kubenswrapper[4989]: I1213 17:31:05.143497 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" Dec 13 17:31:05 crc kubenswrapper[4989]: I1213 17:31:05.143523 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" event={"ID":"0fa9d119-2c72-4737-9eb7-8de2b492f714","Type":"ContainerDied","Data":"7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88"} Dec 13 17:31:05 crc kubenswrapper[4989]: I1213 17:31:05.143589 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-wtqpt" event={"ID":"0fa9d119-2c72-4737-9eb7-8de2b492f714","Type":"ContainerDied","Data":"3a2680737ea99929e54b64fd00668b3d2f368fee98375438086bc6d56e3e4f9b"} Dec 13 17:31:05 crc kubenswrapper[4989]: I1213 17:31:05.143626 4989 scope.go:117] "RemoveContainer" containerID="7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88" Dec 13 17:31:05 crc kubenswrapper[4989]: I1213 17:31:05.163567 4989 scope.go:117] "RemoveContainer" containerID="7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88" Dec 13 17:31:05 crc kubenswrapper[4989]: E1213 17:31:05.164121 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88\": container with ID starting with 7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88 not found: ID does not exist" containerID="7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88" Dec 13 17:31:05 crc kubenswrapper[4989]: I1213 17:31:05.164152 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88"} err="failed to get container status \"7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88\": rpc error: code = NotFound desc = could not find container \"7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88\": container with ID starting with 7037773d951078067297dea4069bb56fbbc3a606986b116eac89ceee85cc1b88 not found: ID does not exist" Dec 13 17:31:05 crc kubenswrapper[4989]: I1213 17:31:05.179662 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-wtqpt"] Dec 13 17:31:05 crc kubenswrapper[4989]: I1213 17:31:05.187583 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-wtqpt"] Dec 13 17:31:05 crc kubenswrapper[4989]: I1213 17:31:05.414468 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:31:05 crc kubenswrapper[4989]: I1213 17:31:05.514036 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:31:06 crc kubenswrapper[4989]: I1213 17:31:06.021722 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fa9d119-2c72-4737-9eb7-8de2b492f714" path="/var/lib/kubelet/pods/0fa9d119-2c72-4737-9eb7-8de2b492f714/volumes" Dec 13 17:31:09 crc kubenswrapper[4989]: I1213 17:31:09.514531 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/memcached-0" Dec 13 17:31:10 crc kubenswrapper[4989]: I1213 17:31:09.845246 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" Dec 13 17:31:10 crc kubenswrapper[4989]: I1213 17:31:09.845381 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" Dec 13 17:31:10 crc kubenswrapper[4989]: I1213 17:31:09.870112 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" Dec 13 17:31:10 crc kubenswrapper[4989]: I1213 17:31:10.200556 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.224342 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/root-account-create-update-hllmp"] Dec 13 17:31:11 crc kubenswrapper[4989]: E1213 17:31:11.224581 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa9d119-2c72-4737-9eb7-8de2b492f714" containerName="registry-server" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.224596 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa9d119-2c72-4737-9eb7-8de2b492f714" containerName="registry-server" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.224715 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fa9d119-2c72-4737-9eb7-8de2b492f714" containerName="registry-server" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.225130 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-hllmp" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.227153 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"openstack-mariadb-root-db-secret" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.240181 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-hllmp"] Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.298097 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk76m\" (UniqueName: \"kubernetes.io/projected/74cceb6b-b34b-42c2-8fe4-b93433e758d0-kube-api-access-lk76m\") pod \"root-account-create-update-hllmp\" (UID: \"74cceb6b-b34b-42c2-8fe4-b93433e758d0\") " pod="cinder-kuttl-tests/root-account-create-update-hllmp" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.298168 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74cceb6b-b34b-42c2-8fe4-b93433e758d0-operator-scripts\") pod \"root-account-create-update-hllmp\" (UID: \"74cceb6b-b34b-42c2-8fe4-b93433e758d0\") " pod="cinder-kuttl-tests/root-account-create-update-hllmp" Dec 13 17:31:11 crc kubenswrapper[4989]: E1213 17:31:11.363852 4989 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.47:47534->38.102.83.47:34293: write tcp 38.102.83.47:47534->38.102.83.47:34293: write: broken pipe Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.399196 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk76m\" (UniqueName: \"kubernetes.io/projected/74cceb6b-b34b-42c2-8fe4-b93433e758d0-kube-api-access-lk76m\") pod \"root-account-create-update-hllmp\" (UID: \"74cceb6b-b34b-42c2-8fe4-b93433e758d0\") " pod="cinder-kuttl-tests/root-account-create-update-hllmp" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.400023 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74cceb6b-b34b-42c2-8fe4-b93433e758d0-operator-scripts\") pod \"root-account-create-update-hllmp\" (UID: \"74cceb6b-b34b-42c2-8fe4-b93433e758d0\") " pod="cinder-kuttl-tests/root-account-create-update-hllmp" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.401037 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74cceb6b-b34b-42c2-8fe4-b93433e758d0-operator-scripts\") pod \"root-account-create-update-hllmp\" (UID: \"74cceb6b-b34b-42c2-8fe4-b93433e758d0\") " pod="cinder-kuttl-tests/root-account-create-update-hllmp" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.425716 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk76m\" (UniqueName: \"kubernetes.io/projected/74cceb6b-b34b-42c2-8fe4-b93433e758d0-kube-api-access-lk76m\") pod \"root-account-create-update-hllmp\" (UID: \"74cceb6b-b34b-42c2-8fe4-b93433e758d0\") " pod="cinder-kuttl-tests/root-account-create-update-hllmp" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.542561 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-hllmp" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.554881 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm"] Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.556160 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.558661 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k762t" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.568499 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm"] Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.602417 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kchx9\" (UniqueName: \"kubernetes.io/projected/5ac1af50-c321-49bf-9270-21c2a4afea6a-kube-api-access-kchx9\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.602689 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.602826 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.704344 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.704431 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.704499 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kchx9\" (UniqueName: \"kubernetes.io/projected/5ac1af50-c321-49bf-9270-21c2a4afea6a-kube-api-access-kchx9\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.705007 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.705006 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.726050 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kchx9\" (UniqueName: \"kubernetes.io/projected/5ac1af50-c321-49bf-9270-21c2a4afea6a-kube-api-access-kchx9\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:11 crc kubenswrapper[4989]: I1213 17:31:11.872087 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:12 crc kubenswrapper[4989]: I1213 17:31:12.973505 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/openstack-galera-2" podUID="c540cd7f-4b50-4a71-9e51-0687156ee1b2" containerName="galera" probeResult="failure" output=< Dec 13 17:31:12 crc kubenswrapper[4989]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Dec 13 17:31:12 crc kubenswrapper[4989]: > Dec 13 17:31:13 crc kubenswrapper[4989]: W1213 17:31:13.388553 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74cceb6b_b34b_42c2_8fe4_b93433e758d0.slice/crio-e19b24bdeae9d2969455b6b85910591b365f8e937d8aad68abd0cb268baafcb3 WatchSource:0}: Error finding container e19b24bdeae9d2969455b6b85910591b365f8e937d8aad68abd0cb268baafcb3: Status 404 returned error can't find the container with id e19b24bdeae9d2969455b6b85910591b365f8e937d8aad68abd0cb268baafcb3 Dec 13 17:31:13 crc kubenswrapper[4989]: I1213 17:31:13.390023 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-hllmp"] Dec 13 17:31:13 crc kubenswrapper[4989]: I1213 17:31:13.433817 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm"] Dec 13 17:31:13 crc kubenswrapper[4989]: W1213 17:31:13.440436 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ac1af50_c321_49bf_9270_21c2a4afea6a.slice/crio-97905f1e6b5dc498357a392157f56885ad324e192b9917b033971f01aa1219c4 WatchSource:0}: Error finding container 97905f1e6b5dc498357a392157f56885ad324e192b9917b033971f01aa1219c4: Status 404 returned error can't find the container with id 97905f1e6b5dc498357a392157f56885ad324e192b9917b033971f01aa1219c4 Dec 13 17:31:14 crc kubenswrapper[4989]: I1213 17:31:14.219679 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-hllmp" event={"ID":"74cceb6b-b34b-42c2-8fe4-b93433e758d0","Type":"ContainerStarted","Data":"e19b24bdeae9d2969455b6b85910591b365f8e937d8aad68abd0cb268baafcb3"} Dec 13 17:31:14 crc kubenswrapper[4989]: I1213 17:31:14.222179 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" event={"ID":"5ac1af50-c321-49bf-9270-21c2a4afea6a","Type":"ContainerStarted","Data":"97905f1e6b5dc498357a392157f56885ad324e192b9917b033971f01aa1219c4"} Dec 13 17:31:16 crc kubenswrapper[4989]: I1213 17:31:16.240254 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-hllmp" event={"ID":"74cceb6b-b34b-42c2-8fe4-b93433e758d0","Type":"ContainerStarted","Data":"424109aa8329e2aa5518d5a8bc6bb03cc68221af0e5ec6a7aec49992552e9603"} Dec 13 17:31:16 crc kubenswrapper[4989]: I1213 17:31:16.246483 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" event={"ID":"5ac1af50-c321-49bf-9270-21c2a4afea6a","Type":"ContainerStarted","Data":"99408b9a55f861ab6c0e10e8c233f9fe5790378eeba2ea3c6f8a001834572670"} Dec 13 17:31:16 crc kubenswrapper[4989]: I1213 17:31:16.266910 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/root-account-create-update-hllmp" podStartSLOduration=5.266882603 podStartE2EDuration="5.266882603s" podCreationTimestamp="2025-12-13 17:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:31:16.263091706 +0000 UTC m=+890.869538844" watchObservedRunningTime="2025-12-13 17:31:16.266882603 +0000 UTC m=+890.873329741" Dec 13 17:31:17 crc kubenswrapper[4989]: I1213 17:31:17.253696 4989 generic.go:334] "Generic (PLEG): container finished" podID="5ac1af50-c321-49bf-9270-21c2a4afea6a" containerID="99408b9a55f861ab6c0e10e8c233f9fe5790378eeba2ea3c6f8a001834572670" exitCode=0 Dec 13 17:31:17 crc kubenswrapper[4989]: I1213 17:31:17.253762 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" event={"ID":"5ac1af50-c321-49bf-9270-21c2a4afea6a","Type":"ContainerDied","Data":"99408b9a55f861ab6c0e10e8c233f9fe5790378eeba2ea3c6f8a001834572670"} Dec 13 17:31:18 crc kubenswrapper[4989]: I1213 17:31:18.270206 4989 generic.go:334] "Generic (PLEG): container finished" podID="74cceb6b-b34b-42c2-8fe4-b93433e758d0" containerID="424109aa8329e2aa5518d5a8bc6bb03cc68221af0e5ec6a7aec49992552e9603" exitCode=0 Dec 13 17:31:18 crc kubenswrapper[4989]: I1213 17:31:18.270271 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-hllmp" event={"ID":"74cceb6b-b34b-42c2-8fe4-b93433e758d0","Type":"ContainerDied","Data":"424109aa8329e2aa5518d5a8bc6bb03cc68221af0e5ec6a7aec49992552e9603"} Dec 13 17:31:19 crc kubenswrapper[4989]: I1213 17:31:19.278861 4989 generic.go:334] "Generic (PLEG): container finished" podID="5ac1af50-c321-49bf-9270-21c2a4afea6a" containerID="69bfdcd39b6722f9908d6c327460719cc72bb5f5c69dda1fbcd0eefeee40325d" exitCode=0 Dec 13 17:31:19 crc kubenswrapper[4989]: I1213 17:31:19.278919 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" event={"ID":"5ac1af50-c321-49bf-9270-21c2a4afea6a","Type":"ContainerDied","Data":"69bfdcd39b6722f9908d6c327460719cc72bb5f5c69dda1fbcd0eefeee40325d"} Dec 13 17:31:19 crc kubenswrapper[4989]: I1213 17:31:19.588278 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-hllmp" Dec 13 17:31:19 crc kubenswrapper[4989]: I1213 17:31:19.728046 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74cceb6b-b34b-42c2-8fe4-b93433e758d0-operator-scripts\") pod \"74cceb6b-b34b-42c2-8fe4-b93433e758d0\" (UID: \"74cceb6b-b34b-42c2-8fe4-b93433e758d0\") " Dec 13 17:31:19 crc kubenswrapper[4989]: I1213 17:31:19.728106 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk76m\" (UniqueName: \"kubernetes.io/projected/74cceb6b-b34b-42c2-8fe4-b93433e758d0-kube-api-access-lk76m\") pod \"74cceb6b-b34b-42c2-8fe4-b93433e758d0\" (UID: \"74cceb6b-b34b-42c2-8fe4-b93433e758d0\") " Dec 13 17:31:19 crc kubenswrapper[4989]: I1213 17:31:19.729105 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74cceb6b-b34b-42c2-8fe4-b93433e758d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "74cceb6b-b34b-42c2-8fe4-b93433e758d0" (UID: "74cceb6b-b34b-42c2-8fe4-b93433e758d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:31:19 crc kubenswrapper[4989]: I1213 17:31:19.740039 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74cceb6b-b34b-42c2-8fe4-b93433e758d0-kube-api-access-lk76m" (OuterVolumeSpecName: "kube-api-access-lk76m") pod "74cceb6b-b34b-42c2-8fe4-b93433e758d0" (UID: "74cceb6b-b34b-42c2-8fe4-b93433e758d0"). InnerVolumeSpecName "kube-api-access-lk76m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:31:19 crc kubenswrapper[4989]: I1213 17:31:19.830023 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/74cceb6b-b34b-42c2-8fe4-b93433e758d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:31:19 crc kubenswrapper[4989]: I1213 17:31:19.830059 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk76m\" (UniqueName: \"kubernetes.io/projected/74cceb6b-b34b-42c2-8fe4-b93433e758d0-kube-api-access-lk76m\") on node \"crc\" DevicePath \"\"" Dec 13 17:31:20 crc kubenswrapper[4989]: I1213 17:31:20.286339 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-hllmp" event={"ID":"74cceb6b-b34b-42c2-8fe4-b93433e758d0","Type":"ContainerDied","Data":"e19b24bdeae9d2969455b6b85910591b365f8e937d8aad68abd0cb268baafcb3"} Dec 13 17:31:20 crc kubenswrapper[4989]: I1213 17:31:20.287180 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e19b24bdeae9d2969455b6b85910591b365f8e937d8aad68abd0cb268baafcb3" Dec 13 17:31:20 crc kubenswrapper[4989]: I1213 17:31:20.286407 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-hllmp" Dec 13 17:31:21 crc kubenswrapper[4989]: I1213 17:31:21.002055 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:31:21 crc kubenswrapper[4989]: I1213 17:31:21.082747 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:31:21 crc kubenswrapper[4989]: I1213 17:31:21.294388 4989 generic.go:334] "Generic (PLEG): container finished" podID="5ac1af50-c321-49bf-9270-21c2a4afea6a" containerID="558803f067315776cf7ac1b786343f01938f94576665abad2bc8c725fdfee1f9" exitCode=0 Dec 13 17:31:21 crc kubenswrapper[4989]: I1213 17:31:21.294426 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" event={"ID":"5ac1af50-c321-49bf-9270-21c2a4afea6a","Type":"ContainerDied","Data":"558803f067315776cf7ac1b786343f01938f94576665abad2bc8c725fdfee1f9"} Dec 13 17:31:22 crc kubenswrapper[4989]: I1213 17:31:22.662644 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:22 crc kubenswrapper[4989]: I1213 17:31:22.767938 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kchx9\" (UniqueName: \"kubernetes.io/projected/5ac1af50-c321-49bf-9270-21c2a4afea6a-kube-api-access-kchx9\") pod \"5ac1af50-c321-49bf-9270-21c2a4afea6a\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " Dec 13 17:31:22 crc kubenswrapper[4989]: I1213 17:31:22.768007 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-util\") pod \"5ac1af50-c321-49bf-9270-21c2a4afea6a\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " Dec 13 17:31:22 crc kubenswrapper[4989]: I1213 17:31:22.768102 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-bundle\") pod \"5ac1af50-c321-49bf-9270-21c2a4afea6a\" (UID: \"5ac1af50-c321-49bf-9270-21c2a4afea6a\") " Dec 13 17:31:22 crc kubenswrapper[4989]: I1213 17:31:22.769032 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-bundle" (OuterVolumeSpecName: "bundle") pod "5ac1af50-c321-49bf-9270-21c2a4afea6a" (UID: "5ac1af50-c321-49bf-9270-21c2a4afea6a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:31:22 crc kubenswrapper[4989]: I1213 17:31:22.779464 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-util" (OuterVolumeSpecName: "util") pod "5ac1af50-c321-49bf-9270-21c2a4afea6a" (UID: "5ac1af50-c321-49bf-9270-21c2a4afea6a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:31:22 crc kubenswrapper[4989]: I1213 17:31:22.781986 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ac1af50-c321-49bf-9270-21c2a4afea6a-kube-api-access-kchx9" (OuterVolumeSpecName: "kube-api-access-kchx9") pod "5ac1af50-c321-49bf-9270-21c2a4afea6a" (UID: "5ac1af50-c321-49bf-9270-21c2a4afea6a"). InnerVolumeSpecName "kube-api-access-kchx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:31:22 crc kubenswrapper[4989]: I1213 17:31:22.869833 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kchx9\" (UniqueName: \"kubernetes.io/projected/5ac1af50-c321-49bf-9270-21c2a4afea6a-kube-api-access-kchx9\") on node \"crc\" DevicePath \"\"" Dec 13 17:31:22 crc kubenswrapper[4989]: I1213 17:31:22.869878 4989 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-util\") on node \"crc\" DevicePath \"\"" Dec 13 17:31:22 crc kubenswrapper[4989]: I1213 17:31:22.869895 4989 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ac1af50-c321-49bf-9270-21c2a4afea6a-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:31:23 crc kubenswrapper[4989]: I1213 17:31:23.308289 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" event={"ID":"5ac1af50-c321-49bf-9270-21c2a4afea6a","Type":"ContainerDied","Data":"97905f1e6b5dc498357a392157f56885ad324e192b9917b033971f01aa1219c4"} Dec 13 17:31:23 crc kubenswrapper[4989]: I1213 17:31:23.308326 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97905f1e6b5dc498357a392157f56885ad324e192b9917b033971f01aa1219c4" Dec 13 17:31:23 crc kubenswrapper[4989]: I1213 17:31:23.308360 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm" Dec 13 17:31:24 crc kubenswrapper[4989]: I1213 17:31:24.401175 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:31:24 crc kubenswrapper[4989]: I1213 17:31:24.502487 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.440998 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg"] Dec 13 17:31:31 crc kubenswrapper[4989]: E1213 17:31:31.442627 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ac1af50-c321-49bf-9270-21c2a4afea6a" containerName="extract" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.442744 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ac1af50-c321-49bf-9270-21c2a4afea6a" containerName="extract" Dec 13 17:31:31 crc kubenswrapper[4989]: E1213 17:31:31.442847 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ac1af50-c321-49bf-9270-21c2a4afea6a" containerName="util" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.442916 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ac1af50-c321-49bf-9270-21c2a4afea6a" containerName="util" Dec 13 17:31:31 crc kubenswrapper[4989]: E1213 17:31:31.443001 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74cceb6b-b34b-42c2-8fe4-b93433e758d0" containerName="mariadb-account-create-update" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.443069 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="74cceb6b-b34b-42c2-8fe4-b93433e758d0" containerName="mariadb-account-create-update" Dec 13 17:31:31 crc kubenswrapper[4989]: E1213 17:31:31.443137 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ac1af50-c321-49bf-9270-21c2a4afea6a" containerName="pull" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.443204 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ac1af50-c321-49bf-9270-21c2a4afea6a" containerName="pull" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.443416 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="74cceb6b-b34b-42c2-8fe4-b93433e758d0" containerName="mariadb-account-create-update" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.443511 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ac1af50-c321-49bf-9270-21c2a4afea6a" containerName="extract" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.444136 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.446146 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-kv4hr" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.453284 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg"] Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.489245 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjvbs\" (UniqueName: \"kubernetes.io/projected/7334147d-9cc6-47cc-b164-56a971e3e82b-kube-api-access-jjvbs\") pod \"rabbitmq-cluster-operator-779fc9694b-k55sg\" (UID: \"7334147d-9cc6-47cc-b164-56a971e3e82b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.590297 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjvbs\" (UniqueName: \"kubernetes.io/projected/7334147d-9cc6-47cc-b164-56a971e3e82b-kube-api-access-jjvbs\") pod \"rabbitmq-cluster-operator-779fc9694b-k55sg\" (UID: \"7334147d-9cc6-47cc-b164-56a971e3e82b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.616738 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjvbs\" (UniqueName: \"kubernetes.io/projected/7334147d-9cc6-47cc-b164-56a971e3e82b-kube-api-access-jjvbs\") pod \"rabbitmq-cluster-operator-779fc9694b-k55sg\" (UID: \"7334147d-9cc6-47cc-b164-56a971e3e82b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" Dec 13 17:31:31 crc kubenswrapper[4989]: I1213 17:31:31.760446 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" Dec 13 17:31:32 crc kubenswrapper[4989]: I1213 17:31:32.293708 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg"] Dec 13 17:31:32 crc kubenswrapper[4989]: I1213 17:31:32.359589 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" event={"ID":"7334147d-9cc6-47cc-b164-56a971e3e82b","Type":"ContainerStarted","Data":"96777a515a4f3d1198bbc0ba3306b85398abb5e4768f1761918d90f097e22088"} Dec 13 17:31:36 crc kubenswrapper[4989]: I1213 17:31:36.392538 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" event={"ID":"7334147d-9cc6-47cc-b164-56a971e3e82b","Type":"ContainerStarted","Data":"b448f0bd85df0ca5d9a877c2a09c90416ca3aca7d5fa9bb8c8e23c6be64bfb93"} Dec 13 17:31:36 crc kubenswrapper[4989]: I1213 17:31:36.421210 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" podStartSLOduration=1.742884205 podStartE2EDuration="5.421185943s" podCreationTimestamp="2025-12-13 17:31:31 +0000 UTC" firstStartedPulling="2025-12-13 17:31:32.300533517 +0000 UTC m=+906.906980655" lastFinishedPulling="2025-12-13 17:31:35.978835255 +0000 UTC m=+910.585282393" observedRunningTime="2025-12-13 17:31:36.412302549 +0000 UTC m=+911.018749697" watchObservedRunningTime="2025-12-13 17:31:36.421185943 +0000 UTC m=+911.027633101" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.221735 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.223240 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.225723 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-erlang-cookie" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.225914 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"rabbitmq-server-conf" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.226000 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-default-user" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.228490 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-server-dockercfg-jfgtt" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.230062 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.230995 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"rabbitmq-plugins-conf" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.365668 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzctk\" (UniqueName: \"kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-kube-api-access-rzctk\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.366218 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78da11d7-96b7-4cfd-b190-f34622427b21-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.366284 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.366328 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.366354 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78da11d7-96b7-4cfd-b190-f34622427b21-pod-info\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.366379 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.366419 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78da11d7-96b7-4cfd-b190-f34622427b21-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.366464 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.468066 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzctk\" (UniqueName: \"kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-kube-api-access-rzctk\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.468121 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78da11d7-96b7-4cfd-b190-f34622427b21-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.468153 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.468185 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78da11d7-96b7-4cfd-b190-f34622427b21-pod-info\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.468208 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.468227 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.468262 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78da11d7-96b7-4cfd-b190-f34622427b21-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.468297 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.468755 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.469712 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78da11d7-96b7-4cfd-b190-f34622427b21-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.471487 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.477313 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.477610 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78da11d7-96b7-4cfd-b190-f34622427b21-pod-info\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.481754 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78da11d7-96b7-4cfd-b190-f34622427b21-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.484833 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzctk\" (UniqueName: \"kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-kube-api-access-rzctk\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.494367 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.494406 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5a6e24df0e146639886e579853f556a7732f71e76e13980c2da2e92830a600a1/globalmount\"" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.534323 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\") pod \"rabbitmq-server-0\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:42 crc kubenswrapper[4989]: I1213 17:31:42.539621 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:31:43 crc kubenswrapper[4989]: I1213 17:31:43.039489 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 13 17:31:43 crc kubenswrapper[4989]: I1213 17:31:43.438894 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"78da11d7-96b7-4cfd-b190-f34622427b21","Type":"ContainerStarted","Data":"73b5371443f3cb2d3401456c60cbb5b79e6a6c8cd3216369158aee6396f71e35"} Dec 13 17:31:43 crc kubenswrapper[4989]: I1213 17:31:43.925995 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-9x9cg"] Dec 13 17:31:43 crc kubenswrapper[4989]: I1213 17:31:43.926734 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-9x9cg" Dec 13 17:31:43 crc kubenswrapper[4989]: I1213 17:31:43.928832 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-j4dd9" Dec 13 17:31:43 crc kubenswrapper[4989]: I1213 17:31:43.936515 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-9x9cg"] Dec 13 17:31:43 crc kubenswrapper[4989]: I1213 17:31:43.991905 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpbvp\" (UniqueName: \"kubernetes.io/projected/c8d4b0de-4254-46f6-99dd-2e077312b011-kube-api-access-fpbvp\") pod \"keystone-operator-index-9x9cg\" (UID: \"c8d4b0de-4254-46f6-99dd-2e077312b011\") " pod="openstack-operators/keystone-operator-index-9x9cg" Dec 13 17:31:44 crc kubenswrapper[4989]: I1213 17:31:44.096603 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpbvp\" (UniqueName: \"kubernetes.io/projected/c8d4b0de-4254-46f6-99dd-2e077312b011-kube-api-access-fpbvp\") pod \"keystone-operator-index-9x9cg\" (UID: \"c8d4b0de-4254-46f6-99dd-2e077312b011\") " pod="openstack-operators/keystone-operator-index-9x9cg" Dec 13 17:31:44 crc kubenswrapper[4989]: I1213 17:31:44.118363 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpbvp\" (UniqueName: \"kubernetes.io/projected/c8d4b0de-4254-46f6-99dd-2e077312b011-kube-api-access-fpbvp\") pod \"keystone-operator-index-9x9cg\" (UID: \"c8d4b0de-4254-46f6-99dd-2e077312b011\") " pod="openstack-operators/keystone-operator-index-9x9cg" Dec 13 17:31:44 crc kubenswrapper[4989]: I1213 17:31:44.259690 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-9x9cg" Dec 13 17:31:44 crc kubenswrapper[4989]: I1213 17:31:44.711114 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-9x9cg"] Dec 13 17:31:44 crc kubenswrapper[4989]: W1213 17:31:44.717885 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8d4b0de_4254_46f6_99dd_2e077312b011.slice/crio-080aad60729212c42dc5aa3472ba223493e9aa0fd612954addccd7ca13432aca WatchSource:0}: Error finding container 080aad60729212c42dc5aa3472ba223493e9aa0fd612954addccd7ca13432aca: Status 404 returned error can't find the container with id 080aad60729212c42dc5aa3472ba223493e9aa0fd612954addccd7ca13432aca Dec 13 17:31:45 crc kubenswrapper[4989]: I1213 17:31:45.451967 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-9x9cg" event={"ID":"c8d4b0de-4254-46f6-99dd-2e077312b011","Type":"ContainerStarted","Data":"080aad60729212c42dc5aa3472ba223493e9aa0fd612954addccd7ca13432aca"} Dec 13 17:31:51 crc kubenswrapper[4989]: I1213 17:31:51.868735 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-9x9cg" event={"ID":"c8d4b0de-4254-46f6-99dd-2e077312b011","Type":"ContainerStarted","Data":"1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1"} Dec 13 17:31:51 crc kubenswrapper[4989]: I1213 17:31:51.886217 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-9x9cg" podStartSLOduration=2.618445296 podStartE2EDuration="8.886199168s" podCreationTimestamp="2025-12-13 17:31:43 +0000 UTC" firstStartedPulling="2025-12-13 17:31:44.720695056 +0000 UTC m=+919.327142194" lastFinishedPulling="2025-12-13 17:31:50.988448928 +0000 UTC m=+925.594896066" observedRunningTime="2025-12-13 17:31:51.882852164 +0000 UTC m=+926.489299342" watchObservedRunningTime="2025-12-13 17:31:51.886199168 +0000 UTC m=+926.492646306" Dec 13 17:31:52 crc kubenswrapper[4989]: I1213 17:31:52.876712 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"78da11d7-96b7-4cfd-b190-f34622427b21","Type":"ContainerStarted","Data":"c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622"} Dec 13 17:31:54 crc kubenswrapper[4989]: I1213 17:31:54.259919 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-9x9cg" Dec 13 17:31:54 crc kubenswrapper[4989]: I1213 17:31:54.259990 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-9x9cg" Dec 13 17:31:54 crc kubenswrapper[4989]: I1213 17:31:54.309936 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-9x9cg" Dec 13 17:32:04 crc kubenswrapper[4989]: I1213 17:32:04.294687 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-9x9cg" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.363854 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx"] Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.365888 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.368348 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k762t" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.375319 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx"] Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.472336 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64mc4\" (UniqueName: \"kubernetes.io/projected/4146fa85-1ad6-492e-b6a8-71db9e308953-kube-api-access-64mc4\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.472411 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-util\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.472479 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-bundle\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.574085 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-bundle\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.574422 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64mc4\" (UniqueName: \"kubernetes.io/projected/4146fa85-1ad6-492e-b6a8-71db9e308953-kube-api-access-64mc4\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.574509 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-util\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.575567 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-bundle\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.575596 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-util\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.596305 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64mc4\" (UniqueName: \"kubernetes.io/projected/4146fa85-1ad6-492e-b6a8-71db9e308953-kube-api-access-64mc4\") pod \"835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:14 crc kubenswrapper[4989]: I1213 17:32:14.706123 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:15 crc kubenswrapper[4989]: I1213 17:32:15.130810 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx"] Dec 13 17:32:16 crc kubenswrapper[4989]: I1213 17:32:16.084190 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" event={"ID":"4146fa85-1ad6-492e-b6a8-71db9e308953","Type":"ContainerStarted","Data":"a633f2e2308fd45a01cc7c159d9db6ce55e2aaf50aa8ca3fe0e977009936ff17"} Dec 13 17:32:17 crc kubenswrapper[4989]: I1213 17:32:17.090150 4989 generic.go:334] "Generic (PLEG): container finished" podID="4146fa85-1ad6-492e-b6a8-71db9e308953" containerID="3ecc834249e0941b8d5a34b4963ecec7ce2f4115f753e77b0bd0be0a8677231a" exitCode=0 Dec 13 17:32:17 crc kubenswrapper[4989]: I1213 17:32:17.090189 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" event={"ID":"4146fa85-1ad6-492e-b6a8-71db9e308953","Type":"ContainerDied","Data":"3ecc834249e0941b8d5a34b4963ecec7ce2f4115f753e77b0bd0be0a8677231a"} Dec 13 17:32:18 crc kubenswrapper[4989]: E1213 17:32:18.086540 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4146fa85_1ad6_492e_b6a8_71db9e308953.slice/crio-conmon-83c90348a987db1c060ab96646093a76fd802d23c8b03282aad61fe4e10bac2a.scope\": RecentStats: unable to find data in memory cache]" Dec 13 17:32:18 crc kubenswrapper[4989]: I1213 17:32:18.098315 4989 generic.go:334] "Generic (PLEG): container finished" podID="4146fa85-1ad6-492e-b6a8-71db9e308953" containerID="83c90348a987db1c060ab96646093a76fd802d23c8b03282aad61fe4e10bac2a" exitCode=0 Dec 13 17:32:18 crc kubenswrapper[4989]: I1213 17:32:18.098353 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" event={"ID":"4146fa85-1ad6-492e-b6a8-71db9e308953","Type":"ContainerDied","Data":"83c90348a987db1c060ab96646093a76fd802d23c8b03282aad61fe4e10bac2a"} Dec 13 17:32:19 crc kubenswrapper[4989]: I1213 17:32:19.108745 4989 generic.go:334] "Generic (PLEG): container finished" podID="4146fa85-1ad6-492e-b6a8-71db9e308953" containerID="d79bbdd009a893187a04a7dc41238e9bc4c5cd0893bcef281570e76f4720e0cc" exitCode=0 Dec 13 17:32:19 crc kubenswrapper[4989]: I1213 17:32:19.108844 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" event={"ID":"4146fa85-1ad6-492e-b6a8-71db9e308953","Type":"ContainerDied","Data":"d79bbdd009a893187a04a7dc41238e9bc4c5cd0893bcef281570e76f4720e0cc"} Dec 13 17:32:20 crc kubenswrapper[4989]: I1213 17:32:20.405745 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:20 crc kubenswrapper[4989]: I1213 17:32:20.561567 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-util\") pod \"4146fa85-1ad6-492e-b6a8-71db9e308953\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " Dec 13 17:32:20 crc kubenswrapper[4989]: I1213 17:32:20.561922 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-bundle\") pod \"4146fa85-1ad6-492e-b6a8-71db9e308953\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " Dec 13 17:32:20 crc kubenswrapper[4989]: I1213 17:32:20.562195 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64mc4\" (UniqueName: \"kubernetes.io/projected/4146fa85-1ad6-492e-b6a8-71db9e308953-kube-api-access-64mc4\") pod \"4146fa85-1ad6-492e-b6a8-71db9e308953\" (UID: \"4146fa85-1ad6-492e-b6a8-71db9e308953\") " Dec 13 17:32:20 crc kubenswrapper[4989]: I1213 17:32:20.562714 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-bundle" (OuterVolumeSpecName: "bundle") pod "4146fa85-1ad6-492e-b6a8-71db9e308953" (UID: "4146fa85-1ad6-492e-b6a8-71db9e308953"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:32:20 crc kubenswrapper[4989]: I1213 17:32:20.572969 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4146fa85-1ad6-492e-b6a8-71db9e308953-kube-api-access-64mc4" (OuterVolumeSpecName: "kube-api-access-64mc4") pod "4146fa85-1ad6-492e-b6a8-71db9e308953" (UID: "4146fa85-1ad6-492e-b6a8-71db9e308953"). InnerVolumeSpecName "kube-api-access-64mc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:32:20 crc kubenswrapper[4989]: I1213 17:32:20.582010 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-util" (OuterVolumeSpecName: "util") pod "4146fa85-1ad6-492e-b6a8-71db9e308953" (UID: "4146fa85-1ad6-492e-b6a8-71db9e308953"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:32:20 crc kubenswrapper[4989]: I1213 17:32:20.663643 4989 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:32:20 crc kubenswrapper[4989]: I1213 17:32:20.663681 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64mc4\" (UniqueName: \"kubernetes.io/projected/4146fa85-1ad6-492e-b6a8-71db9e308953-kube-api-access-64mc4\") on node \"crc\" DevicePath \"\"" Dec 13 17:32:20 crc kubenswrapper[4989]: I1213 17:32:20.663695 4989 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4146fa85-1ad6-492e-b6a8-71db9e308953-util\") on node \"crc\" DevicePath \"\"" Dec 13 17:32:21 crc kubenswrapper[4989]: I1213 17:32:21.125589 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" event={"ID":"4146fa85-1ad6-492e-b6a8-71db9e308953","Type":"ContainerDied","Data":"a633f2e2308fd45a01cc7c159d9db6ce55e2aaf50aa8ca3fe0e977009936ff17"} Dec 13 17:32:21 crc kubenswrapper[4989]: I1213 17:32:21.125636 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a633f2e2308fd45a01cc7c159d9db6ce55e2aaf50aa8ca3fe0e977009936ff17" Dec 13 17:32:21 crc kubenswrapper[4989]: I1213 17:32:21.125660 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx" Dec 13 17:32:24 crc kubenswrapper[4989]: I1213 17:32:24.148071 4989 generic.go:334] "Generic (PLEG): container finished" podID="78da11d7-96b7-4cfd-b190-f34622427b21" containerID="c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622" exitCode=0 Dec 13 17:32:24 crc kubenswrapper[4989]: I1213 17:32:24.148150 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"78da11d7-96b7-4cfd-b190-f34622427b21","Type":"ContainerDied","Data":"c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622"} Dec 13 17:32:25 crc kubenswrapper[4989]: I1213 17:32:25.157845 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"78da11d7-96b7-4cfd-b190-f34622427b21","Type":"ContainerStarted","Data":"9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb"} Dec 13 17:32:25 crc kubenswrapper[4989]: I1213 17:32:25.158375 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:32:25 crc kubenswrapper[4989]: I1213 17:32:25.182314 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.302024745 podStartE2EDuration="44.182286053s" podCreationTimestamp="2025-12-13 17:31:41 +0000 UTC" firstStartedPulling="2025-12-13 17:31:43.043616819 +0000 UTC m=+917.650063957" lastFinishedPulling="2025-12-13 17:31:50.923878127 +0000 UTC m=+925.530325265" observedRunningTime="2025-12-13 17:32:25.180723385 +0000 UTC m=+959.787170533" watchObservedRunningTime="2025-12-13 17:32:25.182286053 +0000 UTC m=+959.788733201" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.283331 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6757956774-t4phl"] Dec 13 17:32:31 crc kubenswrapper[4989]: E1213 17:32:31.284115 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4146fa85-1ad6-492e-b6a8-71db9e308953" containerName="extract" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.284134 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4146fa85-1ad6-492e-b6a8-71db9e308953" containerName="extract" Dec 13 17:32:31 crc kubenswrapper[4989]: E1213 17:32:31.284150 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4146fa85-1ad6-492e-b6a8-71db9e308953" containerName="pull" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.284159 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4146fa85-1ad6-492e-b6a8-71db9e308953" containerName="pull" Dec 13 17:32:31 crc kubenswrapper[4989]: E1213 17:32:31.284177 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4146fa85-1ad6-492e-b6a8-71db9e308953" containerName="util" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.284187 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4146fa85-1ad6-492e-b6a8-71db9e308953" containerName="util" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.284445 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4146fa85-1ad6-492e-b6a8-71db9e308953" containerName="extract" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.285012 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.286841 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.286846 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-sr7gh" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.303426 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6757956774-t4phl"] Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.423355 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn2cq\" (UniqueName: \"kubernetes.io/projected/61c47c47-cdac-40dc-8644-e2cdbccc560e-kube-api-access-xn2cq\") pod \"keystone-operator-controller-manager-6757956774-t4phl\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.423413 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-apiservice-cert\") pod \"keystone-operator-controller-manager-6757956774-t4phl\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.423503 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-webhook-cert\") pod \"keystone-operator-controller-manager-6757956774-t4phl\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.524950 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-apiservice-cert\") pod \"keystone-operator-controller-manager-6757956774-t4phl\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.525036 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-webhook-cert\") pod \"keystone-operator-controller-manager-6757956774-t4phl\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.525083 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn2cq\" (UniqueName: \"kubernetes.io/projected/61c47c47-cdac-40dc-8644-e2cdbccc560e-kube-api-access-xn2cq\") pod \"keystone-operator-controller-manager-6757956774-t4phl\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.532474 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-webhook-cert\") pod \"keystone-operator-controller-manager-6757956774-t4phl\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.532487 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-apiservice-cert\") pod \"keystone-operator-controller-manager-6757956774-t4phl\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.540137 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn2cq\" (UniqueName: \"kubernetes.io/projected/61c47c47-cdac-40dc-8644-e2cdbccc560e-kube-api-access-xn2cq\") pod \"keystone-operator-controller-manager-6757956774-t4phl\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:31 crc kubenswrapper[4989]: I1213 17:32:31.603279 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:32 crc kubenswrapper[4989]: I1213 17:32:32.014069 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6757956774-t4phl"] Dec 13 17:32:32 crc kubenswrapper[4989]: I1213 17:32:32.201513 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" event={"ID":"61c47c47-cdac-40dc-8644-e2cdbccc560e","Type":"ContainerStarted","Data":"654c23dbcb19cc26f432e022889861dad23d134cf151b3ab6c1196120184f979"} Dec 13 17:32:36 crc kubenswrapper[4989]: I1213 17:32:36.229419 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" event={"ID":"61c47c47-cdac-40dc-8644-e2cdbccc560e","Type":"ContainerStarted","Data":"f643bfc403e39a79d0f6a354f4d18832db63b89a48d82571f80f70ea650083c4"} Dec 13 17:32:36 crc kubenswrapper[4989]: I1213 17:32:36.230040 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:36 crc kubenswrapper[4989]: I1213 17:32:36.245944 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" podStartSLOduration=1.81443405 podStartE2EDuration="5.245925427s" podCreationTimestamp="2025-12-13 17:32:31 +0000 UTC" firstStartedPulling="2025-12-13 17:32:32.022133873 +0000 UTC m=+966.628581051" lastFinishedPulling="2025-12-13 17:32:35.4536253 +0000 UTC m=+970.060072428" observedRunningTime="2025-12-13 17:32:36.242399949 +0000 UTC m=+970.848847087" watchObservedRunningTime="2025-12-13 17:32:36.245925427 +0000 UTC m=+970.852372565" Dec 13 17:32:41 crc kubenswrapper[4989]: I1213 17:32:41.613721 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:32:42 crc kubenswrapper[4989]: I1213 17:32:42.543006 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.581498 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-db-create-88nfm"] Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.582993 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-88nfm" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.589729 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-7987-account-create-update-jtc97"] Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.590958 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.592608 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-db-secret" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.633849 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-88nfm"] Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.641526 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-7987-account-create-update-jtc97"] Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.763481 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/692c8262-bf86-44d0-aa9d-6c03509d3763-operator-scripts\") pod \"keystone-7987-account-create-update-jtc97\" (UID: \"692c8262-bf86-44d0-aa9d-6c03509d3763\") " pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.763620 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7w7d\" (UniqueName: \"kubernetes.io/projected/692c8262-bf86-44d0-aa9d-6c03509d3763-kube-api-access-r7w7d\") pod \"keystone-7987-account-create-update-jtc97\" (UID: \"692c8262-bf86-44d0-aa9d-6c03509d3763\") " pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.763765 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ns4d\" (UniqueName: \"kubernetes.io/projected/dd652c3b-9226-437c-a04e-961498888534-kube-api-access-8ns4d\") pod \"keystone-db-create-88nfm\" (UID: \"dd652c3b-9226-437c-a04e-961498888534\") " pod="cinder-kuttl-tests/keystone-db-create-88nfm" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.763893 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd652c3b-9226-437c-a04e-961498888534-operator-scripts\") pod \"keystone-db-create-88nfm\" (UID: \"dd652c3b-9226-437c-a04e-961498888534\") " pod="cinder-kuttl-tests/keystone-db-create-88nfm" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.865536 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ns4d\" (UniqueName: \"kubernetes.io/projected/dd652c3b-9226-437c-a04e-961498888534-kube-api-access-8ns4d\") pod \"keystone-db-create-88nfm\" (UID: \"dd652c3b-9226-437c-a04e-961498888534\") " pod="cinder-kuttl-tests/keystone-db-create-88nfm" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.865650 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd652c3b-9226-437c-a04e-961498888534-operator-scripts\") pod \"keystone-db-create-88nfm\" (UID: \"dd652c3b-9226-437c-a04e-961498888534\") " pod="cinder-kuttl-tests/keystone-db-create-88nfm" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.865745 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/692c8262-bf86-44d0-aa9d-6c03509d3763-operator-scripts\") pod \"keystone-7987-account-create-update-jtc97\" (UID: \"692c8262-bf86-44d0-aa9d-6c03509d3763\") " pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.865778 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7w7d\" (UniqueName: \"kubernetes.io/projected/692c8262-bf86-44d0-aa9d-6c03509d3763-kube-api-access-r7w7d\") pod \"keystone-7987-account-create-update-jtc97\" (UID: \"692c8262-bf86-44d0-aa9d-6c03509d3763\") " pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.866662 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd652c3b-9226-437c-a04e-961498888534-operator-scripts\") pod \"keystone-db-create-88nfm\" (UID: \"dd652c3b-9226-437c-a04e-961498888534\") " pod="cinder-kuttl-tests/keystone-db-create-88nfm" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.866676 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/692c8262-bf86-44d0-aa9d-6c03509d3763-operator-scripts\") pod \"keystone-7987-account-create-update-jtc97\" (UID: \"692c8262-bf86-44d0-aa9d-6c03509d3763\") " pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.894688 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7w7d\" (UniqueName: \"kubernetes.io/projected/692c8262-bf86-44d0-aa9d-6c03509d3763-kube-api-access-r7w7d\") pod \"keystone-7987-account-create-update-jtc97\" (UID: \"692c8262-bf86-44d0-aa9d-6c03509d3763\") " pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.897124 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ns4d\" (UniqueName: \"kubernetes.io/projected/dd652c3b-9226-437c-a04e-961498888534-kube-api-access-8ns4d\") pod \"keystone-db-create-88nfm\" (UID: \"dd652c3b-9226-437c-a04e-961498888534\") " pod="cinder-kuttl-tests/keystone-db-create-88nfm" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.945657 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-88nfm" Dec 13 17:32:46 crc kubenswrapper[4989]: I1213 17:32:46.966006 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" Dec 13 17:32:47 crc kubenswrapper[4989]: I1213 17:32:47.445897 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-7987-account-create-update-jtc97"] Dec 13 17:32:47 crc kubenswrapper[4989]: W1213 17:32:47.448950 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod692c8262_bf86_44d0_aa9d_6c03509d3763.slice/crio-90c281bd00105f4004727bf3ef75d91bbd9611bbe648b1e556f908e1880e7560 WatchSource:0}: Error finding container 90c281bd00105f4004727bf3ef75d91bbd9611bbe648b1e556f908e1880e7560: Status 404 returned error can't find the container with id 90c281bd00105f4004727bf3ef75d91bbd9611bbe648b1e556f908e1880e7560 Dec 13 17:32:47 crc kubenswrapper[4989]: I1213 17:32:47.516351 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-88nfm"] Dec 13 17:32:47 crc kubenswrapper[4989]: W1213 17:32:47.526253 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd652c3b_9226_437c_a04e_961498888534.slice/crio-a29f7310573f03f62ff32c0530795e2f15b331ee65192eff3ca2ed29d776ed3e WatchSource:0}: Error finding container a29f7310573f03f62ff32c0530795e2f15b331ee65192eff3ca2ed29d776ed3e: Status 404 returned error can't find the container with id a29f7310573f03f62ff32c0530795e2f15b331ee65192eff3ca2ed29d776ed3e Dec 13 17:32:48 crc kubenswrapper[4989]: I1213 17:32:48.307843 4989 generic.go:334] "Generic (PLEG): container finished" podID="692c8262-bf86-44d0-aa9d-6c03509d3763" containerID="5d311671a6cc64730cce91cf40f7dbedf82c270d3fe28bdbd32d6780f50522ed" exitCode=0 Dec 13 17:32:48 crc kubenswrapper[4989]: I1213 17:32:48.307944 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" event={"ID":"692c8262-bf86-44d0-aa9d-6c03509d3763","Type":"ContainerDied","Data":"5d311671a6cc64730cce91cf40f7dbedf82c270d3fe28bdbd32d6780f50522ed"} Dec 13 17:32:48 crc kubenswrapper[4989]: I1213 17:32:48.308199 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" event={"ID":"692c8262-bf86-44d0-aa9d-6c03509d3763","Type":"ContainerStarted","Data":"90c281bd00105f4004727bf3ef75d91bbd9611bbe648b1e556f908e1880e7560"} Dec 13 17:32:48 crc kubenswrapper[4989]: I1213 17:32:48.309857 4989 generic.go:334] "Generic (PLEG): container finished" podID="dd652c3b-9226-437c-a04e-961498888534" containerID="0112ac7b8b36b6378ffce6ae04f4262d5d3787408857526c0bf270dad7889142" exitCode=0 Dec 13 17:32:48 crc kubenswrapper[4989]: I1213 17:32:48.309970 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-88nfm" event={"ID":"dd652c3b-9226-437c-a04e-961498888534","Type":"ContainerDied","Data":"0112ac7b8b36b6378ffce6ae04f4262d5d3787408857526c0bf270dad7889142"} Dec 13 17:32:48 crc kubenswrapper[4989]: I1213 17:32:48.310054 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-88nfm" event={"ID":"dd652c3b-9226-437c-a04e-961498888534","Type":"ContainerStarted","Data":"a29f7310573f03f62ff32c0530795e2f15b331ee65192eff3ca2ed29d776ed3e"} Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.521716 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-index-zjpg6"] Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.530166 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-zjpg6" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.530741 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-index-zjpg6"] Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.532468 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-index-dockercfg-4m657" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.631911 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.636922 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-88nfm" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.710190 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vwkg\" (UniqueName: \"kubernetes.io/projected/f814e529-71ee-4f12-8057-ebfe2cd78b97-kube-api-access-6vwkg\") pod \"cinder-operator-index-zjpg6\" (UID: \"f814e529-71ee-4f12-8057-ebfe2cd78b97\") " pod="openstack-operators/cinder-operator-index-zjpg6" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.811036 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ns4d\" (UniqueName: \"kubernetes.io/projected/dd652c3b-9226-437c-a04e-961498888534-kube-api-access-8ns4d\") pod \"dd652c3b-9226-437c-a04e-961498888534\" (UID: \"dd652c3b-9226-437c-a04e-961498888534\") " Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.811108 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/692c8262-bf86-44d0-aa9d-6c03509d3763-operator-scripts\") pod \"692c8262-bf86-44d0-aa9d-6c03509d3763\" (UID: \"692c8262-bf86-44d0-aa9d-6c03509d3763\") " Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.811172 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7w7d\" (UniqueName: \"kubernetes.io/projected/692c8262-bf86-44d0-aa9d-6c03509d3763-kube-api-access-r7w7d\") pod \"692c8262-bf86-44d0-aa9d-6c03509d3763\" (UID: \"692c8262-bf86-44d0-aa9d-6c03509d3763\") " Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.811255 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd652c3b-9226-437c-a04e-961498888534-operator-scripts\") pod \"dd652c3b-9226-437c-a04e-961498888534\" (UID: \"dd652c3b-9226-437c-a04e-961498888534\") " Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.811478 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vwkg\" (UniqueName: \"kubernetes.io/projected/f814e529-71ee-4f12-8057-ebfe2cd78b97-kube-api-access-6vwkg\") pod \"cinder-operator-index-zjpg6\" (UID: \"f814e529-71ee-4f12-8057-ebfe2cd78b97\") " pod="openstack-operators/cinder-operator-index-zjpg6" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.811854 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/692c8262-bf86-44d0-aa9d-6c03509d3763-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "692c8262-bf86-44d0-aa9d-6c03509d3763" (UID: "692c8262-bf86-44d0-aa9d-6c03509d3763"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.812157 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd652c3b-9226-437c-a04e-961498888534-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd652c3b-9226-437c-a04e-961498888534" (UID: "dd652c3b-9226-437c-a04e-961498888534"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.816423 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/692c8262-bf86-44d0-aa9d-6c03509d3763-kube-api-access-r7w7d" (OuterVolumeSpecName: "kube-api-access-r7w7d") pod "692c8262-bf86-44d0-aa9d-6c03509d3763" (UID: "692c8262-bf86-44d0-aa9d-6c03509d3763"). InnerVolumeSpecName "kube-api-access-r7w7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.816477 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd652c3b-9226-437c-a04e-961498888534-kube-api-access-8ns4d" (OuterVolumeSpecName: "kube-api-access-8ns4d") pod "dd652c3b-9226-437c-a04e-961498888534" (UID: "dd652c3b-9226-437c-a04e-961498888534"). InnerVolumeSpecName "kube-api-access-8ns4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.828865 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vwkg\" (UniqueName: \"kubernetes.io/projected/f814e529-71ee-4f12-8057-ebfe2cd78b97-kube-api-access-6vwkg\") pod \"cinder-operator-index-zjpg6\" (UID: \"f814e529-71ee-4f12-8057-ebfe2cd78b97\") " pod="openstack-operators/cinder-operator-index-zjpg6" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.912580 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd652c3b-9226-437c-a04e-961498888534-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.912611 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ns4d\" (UniqueName: \"kubernetes.io/projected/dd652c3b-9226-437c-a04e-961498888534-kube-api-access-8ns4d\") on node \"crc\" DevicePath \"\"" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.912622 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/692c8262-bf86-44d0-aa9d-6c03509d3763-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.912631 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7w7d\" (UniqueName: \"kubernetes.io/projected/692c8262-bf86-44d0-aa9d-6c03509d3763-kube-api-access-r7w7d\") on node \"crc\" DevicePath \"\"" Dec 13 17:32:49 crc kubenswrapper[4989]: I1213 17:32:49.928576 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-zjpg6" Dec 13 17:32:50 crc kubenswrapper[4989]: I1213 17:32:50.348332 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" event={"ID":"692c8262-bf86-44d0-aa9d-6c03509d3763","Type":"ContainerDied","Data":"90c281bd00105f4004727bf3ef75d91bbd9611bbe648b1e556f908e1880e7560"} Dec 13 17:32:50 crc kubenswrapper[4989]: I1213 17:32:50.348762 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90c281bd00105f4004727bf3ef75d91bbd9611bbe648b1e556f908e1880e7560" Dec 13 17:32:50 crc kubenswrapper[4989]: I1213 17:32:50.348924 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-7987-account-create-update-jtc97" Dec 13 17:32:50 crc kubenswrapper[4989]: I1213 17:32:50.352970 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-88nfm" event={"ID":"dd652c3b-9226-437c-a04e-961498888534","Type":"ContainerDied","Data":"a29f7310573f03f62ff32c0530795e2f15b331ee65192eff3ca2ed29d776ed3e"} Dec 13 17:32:50 crc kubenswrapper[4989]: I1213 17:32:50.353003 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a29f7310573f03f62ff32c0530795e2f15b331ee65192eff3ca2ed29d776ed3e" Dec 13 17:32:50 crc kubenswrapper[4989]: I1213 17:32:50.353050 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-88nfm" Dec 13 17:32:50 crc kubenswrapper[4989]: I1213 17:32:50.387957 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-index-zjpg6"] Dec 13 17:32:51 crc kubenswrapper[4989]: I1213 17:32:51.381063 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-zjpg6" event={"ID":"f814e529-71ee-4f12-8057-ebfe2cd78b97","Type":"ContainerStarted","Data":"b8cca2e7bc17f710d977a79df1bb0ea068c0593096627fc1dba347c8a3524867"} Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.093814 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-7wrn9"] Dec 13 17:32:52 crc kubenswrapper[4989]: E1213 17:32:52.094601 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="692c8262-bf86-44d0-aa9d-6c03509d3763" containerName="mariadb-account-create-update" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.094686 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="692c8262-bf86-44d0-aa9d-6c03509d3763" containerName="mariadb-account-create-update" Dec 13 17:32:52 crc kubenswrapper[4989]: E1213 17:32:52.094765 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd652c3b-9226-437c-a04e-961498888534" containerName="mariadb-database-create" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.094851 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd652c3b-9226-437c-a04e-961498888534" containerName="mariadb-database-create" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.095040 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="692c8262-bf86-44d0-aa9d-6c03509d3763" containerName="mariadb-account-create-update" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.095109 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd652c3b-9226-437c-a04e-961498888534" containerName="mariadb-database-create" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.095695 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.098143 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.098325 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.099321 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.100576 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-8p4xb" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.108544 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-7wrn9"] Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.160781 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhntf\" (UniqueName: \"kubernetes.io/projected/c0587684-6ad8-4723-8fcc-bcb8abbbae22-kube-api-access-nhntf\") pod \"keystone-db-sync-7wrn9\" (UID: \"c0587684-6ad8-4723-8fcc-bcb8abbbae22\") " pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.160877 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0587684-6ad8-4723-8fcc-bcb8abbbae22-config-data\") pod \"keystone-db-sync-7wrn9\" (UID: \"c0587684-6ad8-4723-8fcc-bcb8abbbae22\") " pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.261693 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhntf\" (UniqueName: \"kubernetes.io/projected/c0587684-6ad8-4723-8fcc-bcb8abbbae22-kube-api-access-nhntf\") pod \"keystone-db-sync-7wrn9\" (UID: \"c0587684-6ad8-4723-8fcc-bcb8abbbae22\") " pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.261830 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0587684-6ad8-4723-8fcc-bcb8abbbae22-config-data\") pod \"keystone-db-sync-7wrn9\" (UID: \"c0587684-6ad8-4723-8fcc-bcb8abbbae22\") " pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.268416 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0587684-6ad8-4723-8fcc-bcb8abbbae22-config-data\") pod \"keystone-db-sync-7wrn9\" (UID: \"c0587684-6ad8-4723-8fcc-bcb8abbbae22\") " pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.279894 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhntf\" (UniqueName: \"kubernetes.io/projected/c0587684-6ad8-4723-8fcc-bcb8abbbae22-kube-api-access-nhntf\") pod \"keystone-db-sync-7wrn9\" (UID: \"c0587684-6ad8-4723-8fcc-bcb8abbbae22\") " pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.389058 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-zjpg6" event={"ID":"f814e529-71ee-4f12-8057-ebfe2cd78b97","Type":"ContainerStarted","Data":"0c3ba073df55d68d444ae1a53438f841ae80ba8f57ef35005a2764708a336392"} Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.418534 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-index-zjpg6" podStartSLOduration=1.812697419 podStartE2EDuration="3.418519476s" podCreationTimestamp="2025-12-13 17:32:49 +0000 UTC" firstStartedPulling="2025-12-13 17:32:50.394005635 +0000 UTC m=+985.000452773" lastFinishedPulling="2025-12-13 17:32:51.999827692 +0000 UTC m=+986.606274830" observedRunningTime="2025-12-13 17:32:52.416355499 +0000 UTC m=+987.022802637" watchObservedRunningTime="2025-12-13 17:32:52.418519476 +0000 UTC m=+987.024966614" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.430717 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" Dec 13 17:32:52 crc kubenswrapper[4989]: I1213 17:32:52.865589 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-7wrn9"] Dec 13 17:32:52 crc kubenswrapper[4989]: W1213 17:32:52.868873 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0587684_6ad8_4723_8fcc_bcb8abbbae22.slice/crio-6f14c74992e36c2869420104f0b9061bc460d3681e84ac72c20a15c283ee0730 WatchSource:0}: Error finding container 6f14c74992e36c2869420104f0b9061bc460d3681e84ac72c20a15c283ee0730: Status 404 returned error can't find the container with id 6f14c74992e36c2869420104f0b9061bc460d3681e84ac72c20a15c283ee0730 Dec 13 17:32:53 crc kubenswrapper[4989]: I1213 17:32:53.396250 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" event={"ID":"c0587684-6ad8-4723-8fcc-bcb8abbbae22","Type":"ContainerStarted","Data":"6f14c74992e36c2869420104f0b9061bc460d3681e84ac72c20a15c283ee0730"} Dec 13 17:32:59 crc kubenswrapper[4989]: I1213 17:32:59.928674 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-index-zjpg6" Dec 13 17:32:59 crc kubenswrapper[4989]: I1213 17:32:59.929419 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/cinder-operator-index-zjpg6" Dec 13 17:32:59 crc kubenswrapper[4989]: I1213 17:32:59.964650 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/cinder-operator-index-zjpg6" Dec 13 17:33:00 crc kubenswrapper[4989]: I1213 17:33:00.458071 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" event={"ID":"c0587684-6ad8-4723-8fcc-bcb8abbbae22","Type":"ContainerStarted","Data":"0ac6efa75569331505db0a8eaf9130e6b79234a6680e073a55690ca42ff089fd"} Dec 13 17:33:00 crc kubenswrapper[4989]: I1213 17:33:00.473542 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" podStartSLOduration=1.0694646190000001 podStartE2EDuration="8.473523098s" podCreationTimestamp="2025-12-13 17:32:52 +0000 UTC" firstStartedPulling="2025-12-13 17:32:52.87100342 +0000 UTC m=+987.477450558" lastFinishedPulling="2025-12-13 17:33:00.275061889 +0000 UTC m=+994.881509037" observedRunningTime="2025-12-13 17:33:00.470517355 +0000 UTC m=+995.076964513" watchObservedRunningTime="2025-12-13 17:33:00.473523098 +0000 UTC m=+995.079970236" Dec 13 17:33:00 crc kubenswrapper[4989]: I1213 17:33:00.491123 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-index-zjpg6" Dec 13 17:33:02 crc kubenswrapper[4989]: I1213 17:33:02.765576 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876"] Dec 13 17:33:02 crc kubenswrapper[4989]: I1213 17:33:02.768083 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:02 crc kubenswrapper[4989]: I1213 17:33:02.775039 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k762t" Dec 13 17:33:02 crc kubenswrapper[4989]: I1213 17:33:02.787861 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876"] Dec 13 17:33:02 crc kubenswrapper[4989]: I1213 17:33:02.917071 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-util\") pod \"238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:02 crc kubenswrapper[4989]: I1213 17:33:02.917114 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-bundle\") pod \"238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:02 crc kubenswrapper[4989]: I1213 17:33:02.917139 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2d84\" (UniqueName: \"kubernetes.io/projected/ec03663b-83e2-452f-8229-daa7887134be-kube-api-access-d2d84\") pod \"238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:03 crc kubenswrapper[4989]: I1213 17:33:03.018629 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-util\") pod \"238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:03 crc kubenswrapper[4989]: I1213 17:33:03.018677 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-bundle\") pod \"238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:03 crc kubenswrapper[4989]: I1213 17:33:03.018712 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2d84\" (UniqueName: \"kubernetes.io/projected/ec03663b-83e2-452f-8229-daa7887134be-kube-api-access-d2d84\") pod \"238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:03 crc kubenswrapper[4989]: I1213 17:33:03.019220 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-util\") pod \"238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:03 crc kubenswrapper[4989]: I1213 17:33:03.019236 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-bundle\") pod \"238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:03 crc kubenswrapper[4989]: I1213 17:33:03.042057 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2d84\" (UniqueName: \"kubernetes.io/projected/ec03663b-83e2-452f-8229-daa7887134be-kube-api-access-d2d84\") pod \"238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:03 crc kubenswrapper[4989]: I1213 17:33:03.088438 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:03 crc kubenswrapper[4989]: I1213 17:33:03.404411 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876"] Dec 13 17:33:03 crc kubenswrapper[4989]: W1213 17:33:03.412162 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec03663b_83e2_452f_8229_daa7887134be.slice/crio-8eeca33de82f06b97edd99b76823b1a7a94ddf58df75a9f27f4ca5c478a6dae7 WatchSource:0}: Error finding container 8eeca33de82f06b97edd99b76823b1a7a94ddf58df75a9f27f4ca5c478a6dae7: Status 404 returned error can't find the container with id 8eeca33de82f06b97edd99b76823b1a7a94ddf58df75a9f27f4ca5c478a6dae7 Dec 13 17:33:03 crc kubenswrapper[4989]: I1213 17:33:03.481487 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" event={"ID":"ec03663b-83e2-452f-8229-daa7887134be","Type":"ContainerStarted","Data":"8eeca33de82f06b97edd99b76823b1a7a94ddf58df75a9f27f4ca5c478a6dae7"} Dec 13 17:33:03 crc kubenswrapper[4989]: I1213 17:33:03.501046 4989 generic.go:334] "Generic (PLEG): container finished" podID="c0587684-6ad8-4723-8fcc-bcb8abbbae22" containerID="0ac6efa75569331505db0a8eaf9130e6b79234a6680e073a55690ca42ff089fd" exitCode=0 Dec 13 17:33:03 crc kubenswrapper[4989]: I1213 17:33:03.501130 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" event={"ID":"c0587684-6ad8-4723-8fcc-bcb8abbbae22","Type":"ContainerDied","Data":"0ac6efa75569331505db0a8eaf9130e6b79234a6680e073a55690ca42ff089fd"} Dec 13 17:33:04 crc kubenswrapper[4989]: I1213 17:33:04.510753 4989 generic.go:334] "Generic (PLEG): container finished" podID="ec03663b-83e2-452f-8229-daa7887134be" containerID="4eded5d33a154bfc6fed1cf9d904d9b07585b44d62d7d663caade4661dfdba92" exitCode=0 Dec 13 17:33:04 crc kubenswrapper[4989]: I1213 17:33:04.510877 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" event={"ID":"ec03663b-83e2-452f-8229-daa7887134be","Type":"ContainerDied","Data":"4eded5d33a154bfc6fed1cf9d904d9b07585b44d62d7d663caade4661dfdba92"} Dec 13 17:33:04 crc kubenswrapper[4989]: I1213 17:33:04.514522 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 17:33:04 crc kubenswrapper[4989]: I1213 17:33:04.915614 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.046067 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhntf\" (UniqueName: \"kubernetes.io/projected/c0587684-6ad8-4723-8fcc-bcb8abbbae22-kube-api-access-nhntf\") pod \"c0587684-6ad8-4723-8fcc-bcb8abbbae22\" (UID: \"c0587684-6ad8-4723-8fcc-bcb8abbbae22\") " Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.046426 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0587684-6ad8-4723-8fcc-bcb8abbbae22-config-data\") pod \"c0587684-6ad8-4723-8fcc-bcb8abbbae22\" (UID: \"c0587684-6ad8-4723-8fcc-bcb8abbbae22\") " Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.053056 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0587684-6ad8-4723-8fcc-bcb8abbbae22-kube-api-access-nhntf" (OuterVolumeSpecName: "kube-api-access-nhntf") pod "c0587684-6ad8-4723-8fcc-bcb8abbbae22" (UID: "c0587684-6ad8-4723-8fcc-bcb8abbbae22"). InnerVolumeSpecName "kube-api-access-nhntf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.089708 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0587684-6ad8-4723-8fcc-bcb8abbbae22-config-data" (OuterVolumeSpecName: "config-data") pod "c0587684-6ad8-4723-8fcc-bcb8abbbae22" (UID: "c0587684-6ad8-4723-8fcc-bcb8abbbae22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.148525 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhntf\" (UniqueName: \"kubernetes.io/projected/c0587684-6ad8-4723-8fcc-bcb8abbbae22-kube-api-access-nhntf\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.148559 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0587684-6ad8-4723-8fcc-bcb8abbbae22-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.521943 4989 generic.go:334] "Generic (PLEG): container finished" podID="ec03663b-83e2-452f-8229-daa7887134be" containerID="b204627536e2bd230ccdbce339c40382083b723f0251ec3a5c252ec4509e1e65" exitCode=0 Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.521991 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" event={"ID":"ec03663b-83e2-452f-8229-daa7887134be","Type":"ContainerDied","Data":"b204627536e2bd230ccdbce339c40382083b723f0251ec3a5c252ec4509e1e65"} Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.524027 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" event={"ID":"c0587684-6ad8-4723-8fcc-bcb8abbbae22","Type":"ContainerDied","Data":"6f14c74992e36c2869420104f0b9061bc460d3681e84ac72c20a15c283ee0730"} Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.524072 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f14c74992e36c2869420104f0b9061bc460d3681e84ac72c20a15c283ee0730" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.524159 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-7wrn9" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.706447 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-66jpr"] Dec 13 17:33:05 crc kubenswrapper[4989]: E1213 17:33:05.706902 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0587684-6ad8-4723-8fcc-bcb8abbbae22" containerName="keystone-db-sync" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.706924 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0587684-6ad8-4723-8fcc-bcb8abbbae22" containerName="keystone-db-sync" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.707969 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0587684-6ad8-4723-8fcc-bcb8abbbae22" containerName="keystone-db-sync" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.708651 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.712436 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"osp-secret" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.713610 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.713726 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.714635 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.716203 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-8p4xb" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.717918 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-66jpr"] Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.856467 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h8kp\" (UniqueName: \"kubernetes.io/projected/e61948e3-aa05-40c9-92be-ff2eb7bc9412-kube-api-access-2h8kp\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.856854 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-config-data\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.856890 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-fernet-keys\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.856924 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-credential-keys\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.856939 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-scripts\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.958314 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h8kp\" (UniqueName: \"kubernetes.io/projected/e61948e3-aa05-40c9-92be-ff2eb7bc9412-kube-api-access-2h8kp\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.958423 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-config-data\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.959100 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-fernet-keys\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.959154 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-credential-keys\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.959175 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-scripts\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.964520 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-scripts\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.964889 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-fernet-keys\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.965929 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-config-data\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.965963 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-credential-keys\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:05 crc kubenswrapper[4989]: I1213 17:33:05.978315 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h8kp\" (UniqueName: \"kubernetes.io/projected/e61948e3-aa05-40c9-92be-ff2eb7bc9412-kube-api-access-2h8kp\") pod \"keystone-bootstrap-66jpr\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:06 crc kubenswrapper[4989]: I1213 17:33:06.024059 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:06 crc kubenswrapper[4989]: I1213 17:33:06.488740 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-66jpr"] Dec 13 17:33:06 crc kubenswrapper[4989]: I1213 17:33:06.533591 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" event={"ID":"e61948e3-aa05-40c9-92be-ff2eb7bc9412","Type":"ContainerStarted","Data":"eb9772d94e1a90e885ace60783e71160498a35ececbeb05d72a312e8b565a470"} Dec 13 17:33:06 crc kubenswrapper[4989]: I1213 17:33:06.535879 4989 generic.go:334] "Generic (PLEG): container finished" podID="ec03663b-83e2-452f-8229-daa7887134be" containerID="0be360a14bddf56a685592fbdda141d9a4ccbefb17d572ca98b8362dc89d57df" exitCode=0 Dec 13 17:33:06 crc kubenswrapper[4989]: I1213 17:33:06.535935 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" event={"ID":"ec03663b-83e2-452f-8229-daa7887134be","Type":"ContainerDied","Data":"0be360a14bddf56a685592fbdda141d9a4ccbefb17d572ca98b8362dc89d57df"} Dec 13 17:33:07 crc kubenswrapper[4989]: I1213 17:33:07.548777 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" event={"ID":"e61948e3-aa05-40c9-92be-ff2eb7bc9412","Type":"ContainerStarted","Data":"3bc9eac8dd55bc4b680ff962671d8574d0fa09d888fe7108a89428dab407a2bf"} Dec 13 17:33:07 crc kubenswrapper[4989]: I1213 17:33:07.586988 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" podStartSLOduration=2.586964529 podStartE2EDuration="2.586964529s" podCreationTimestamp="2025-12-13 17:33:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:33:07.585523604 +0000 UTC m=+1002.191970762" watchObservedRunningTime="2025-12-13 17:33:07.586964529 +0000 UTC m=+1002.193411677" Dec 13 17:33:07 crc kubenswrapper[4989]: I1213 17:33:07.950461 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.093927 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2d84\" (UniqueName: \"kubernetes.io/projected/ec03663b-83e2-452f-8229-daa7887134be-kube-api-access-d2d84\") pod \"ec03663b-83e2-452f-8229-daa7887134be\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.094316 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-util\") pod \"ec03663b-83e2-452f-8229-daa7887134be\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.094405 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-bundle\") pod \"ec03663b-83e2-452f-8229-daa7887134be\" (UID: \"ec03663b-83e2-452f-8229-daa7887134be\") " Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.095961 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-bundle" (OuterVolumeSpecName: "bundle") pod "ec03663b-83e2-452f-8229-daa7887134be" (UID: "ec03663b-83e2-452f-8229-daa7887134be"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.102255 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec03663b-83e2-452f-8229-daa7887134be-kube-api-access-d2d84" (OuterVolumeSpecName: "kube-api-access-d2d84") pod "ec03663b-83e2-452f-8229-daa7887134be" (UID: "ec03663b-83e2-452f-8229-daa7887134be"). InnerVolumeSpecName "kube-api-access-d2d84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.128403 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-util" (OuterVolumeSpecName: "util") pod "ec03663b-83e2-452f-8229-daa7887134be" (UID: "ec03663b-83e2-452f-8229-daa7887134be"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.195553 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2d84\" (UniqueName: \"kubernetes.io/projected/ec03663b-83e2-452f-8229-daa7887134be-kube-api-access-d2d84\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.195590 4989 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-util\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.195602 4989 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec03663b-83e2-452f-8229-daa7887134be-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.553889 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" event={"ID":"ec03663b-83e2-452f-8229-daa7887134be","Type":"ContainerDied","Data":"8eeca33de82f06b97edd99b76823b1a7a94ddf58df75a9f27f4ca5c478a6dae7"} Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.553941 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8eeca33de82f06b97edd99b76823b1a7a94ddf58df75a9f27f4ca5c478a6dae7" Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.553960 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876" Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.828991 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:33:08 crc kubenswrapper[4989]: I1213 17:33:08.829054 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:33:10 crc kubenswrapper[4989]: I1213 17:33:10.582964 4989 generic.go:334] "Generic (PLEG): container finished" podID="e61948e3-aa05-40c9-92be-ff2eb7bc9412" containerID="3bc9eac8dd55bc4b680ff962671d8574d0fa09d888fe7108a89428dab407a2bf" exitCode=0 Dec 13 17:33:10 crc kubenswrapper[4989]: I1213 17:33:10.583261 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" event={"ID":"e61948e3-aa05-40c9-92be-ff2eb7bc9412","Type":"ContainerDied","Data":"3bc9eac8dd55bc4b680ff962671d8574d0fa09d888fe7108a89428dab407a2bf"} Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.007160 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.146213 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h8kp\" (UniqueName: \"kubernetes.io/projected/e61948e3-aa05-40c9-92be-ff2eb7bc9412-kube-api-access-2h8kp\") pod \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.146340 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-credential-keys\") pod \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.146385 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-fernet-keys\") pod \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.146408 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-scripts\") pod \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.146472 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-config-data\") pod \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\" (UID: \"e61948e3-aa05-40c9-92be-ff2eb7bc9412\") " Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.151239 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e61948e3-aa05-40c9-92be-ff2eb7bc9412" (UID: "e61948e3-aa05-40c9-92be-ff2eb7bc9412"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.151645 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-scripts" (OuterVolumeSpecName: "scripts") pod "e61948e3-aa05-40c9-92be-ff2eb7bc9412" (UID: "e61948e3-aa05-40c9-92be-ff2eb7bc9412"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.151869 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e61948e3-aa05-40c9-92be-ff2eb7bc9412" (UID: "e61948e3-aa05-40c9-92be-ff2eb7bc9412"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.154430 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e61948e3-aa05-40c9-92be-ff2eb7bc9412-kube-api-access-2h8kp" (OuterVolumeSpecName: "kube-api-access-2h8kp") pod "e61948e3-aa05-40c9-92be-ff2eb7bc9412" (UID: "e61948e3-aa05-40c9-92be-ff2eb7bc9412"). InnerVolumeSpecName "kube-api-access-2h8kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.182690 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-config-data" (OuterVolumeSpecName: "config-data") pod "e61948e3-aa05-40c9-92be-ff2eb7bc9412" (UID: "e61948e3-aa05-40c9-92be-ff2eb7bc9412"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.248252 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.248293 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h8kp\" (UniqueName: \"kubernetes.io/projected/e61948e3-aa05-40c9-92be-ff2eb7bc9412-kube-api-access-2h8kp\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.248308 4989 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.248318 4989 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.248329 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e61948e3-aa05-40c9-92be-ff2eb7bc9412-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.595905 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" event={"ID":"e61948e3-aa05-40c9-92be-ff2eb7bc9412","Type":"ContainerDied","Data":"eb9772d94e1a90e885ace60783e71160498a35ececbeb05d72a312e8b565a470"} Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.595951 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-66jpr" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.595954 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb9772d94e1a90e885ace60783e71160498a35ececbeb05d72a312e8b565a470" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.682164 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-58668b4874-gcpk9"] Dec 13 17:33:12 crc kubenswrapper[4989]: E1213 17:33:12.682434 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e61948e3-aa05-40c9-92be-ff2eb7bc9412" containerName="keystone-bootstrap" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.682451 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e61948e3-aa05-40c9-92be-ff2eb7bc9412" containerName="keystone-bootstrap" Dec 13 17:33:12 crc kubenswrapper[4989]: E1213 17:33:12.682460 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec03663b-83e2-452f-8229-daa7887134be" containerName="util" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.682467 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec03663b-83e2-452f-8229-daa7887134be" containerName="util" Dec 13 17:33:12 crc kubenswrapper[4989]: E1213 17:33:12.682481 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec03663b-83e2-452f-8229-daa7887134be" containerName="pull" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.682487 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec03663b-83e2-452f-8229-daa7887134be" containerName="pull" Dec 13 17:33:12 crc kubenswrapper[4989]: E1213 17:33:12.682496 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec03663b-83e2-452f-8229-daa7887134be" containerName="extract" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.682501 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec03663b-83e2-452f-8229-daa7887134be" containerName="extract" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.682617 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec03663b-83e2-452f-8229-daa7887134be" containerName="extract" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.682627 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e61948e3-aa05-40c9-92be-ff2eb7bc9412" containerName="keystone-bootstrap" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.683122 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.685304 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-8p4xb" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.685582 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.686042 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.686255 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.697109 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-58668b4874-gcpk9"] Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.856778 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-config-data\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.856869 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-fernet-keys\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.856900 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d9pr\" (UniqueName: \"kubernetes.io/projected/69501fce-9a39-435d-a689-6b1eddcf7188-kube-api-access-4d9pr\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.856933 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-scripts\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.856960 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-credential-keys\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.958510 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-fernet-keys\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.958815 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d9pr\" (UniqueName: \"kubernetes.io/projected/69501fce-9a39-435d-a689-6b1eddcf7188-kube-api-access-4d9pr\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.958859 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-scripts\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.958887 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-credential-keys\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.958954 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-config-data\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.964425 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-scripts\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.965119 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-config-data\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.965448 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-fernet-keys\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.966329 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-credential-keys\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.974430 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d9pr\" (UniqueName: \"kubernetes.io/projected/69501fce-9a39-435d-a689-6b1eddcf7188-kube-api-access-4d9pr\") pod \"keystone-58668b4874-gcpk9\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:12 crc kubenswrapper[4989]: I1213 17:33:12.998588 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:13 crc kubenswrapper[4989]: I1213 17:33:13.428583 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-58668b4874-gcpk9"] Dec 13 17:33:13 crc kubenswrapper[4989]: I1213 17:33:13.604492 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" event={"ID":"69501fce-9a39-435d-a689-6b1eddcf7188","Type":"ContainerStarted","Data":"1a2bff07f73140ade3f6b985348f3c9dbed6cf3f6f4a143db52397a2bdc662af"} Dec 13 17:33:14 crc kubenswrapper[4989]: I1213 17:33:14.615007 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" event={"ID":"69501fce-9a39-435d-a689-6b1eddcf7188","Type":"ContainerStarted","Data":"64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852"} Dec 13 17:33:14 crc kubenswrapper[4989]: I1213 17:33:14.615443 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.490070 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" podStartSLOduration=4.490045422 podStartE2EDuration="4.490045422s" podCreationTimestamp="2025-12-13 17:33:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:33:14.643757382 +0000 UTC m=+1009.250204560" watchObservedRunningTime="2025-12-13 17:33:16.490045422 +0000 UTC m=+1011.096492560" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.493230 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv"] Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.494239 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.497058 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-s99k8" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.500480 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-service-cert" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.516754 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv"] Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.619915 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-apiservice-cert\") pod \"cinder-operator-controller-manager-56ddbfcbfc-bclmv\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.619985 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-webhook-cert\") pod \"cinder-operator-controller-manager-56ddbfcbfc-bclmv\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.620128 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bwfh\" (UniqueName: \"kubernetes.io/projected/93cc60c5-709e-474e-81c5-04242bbd0dec-kube-api-access-6bwfh\") pod \"cinder-operator-controller-manager-56ddbfcbfc-bclmv\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.721918 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bwfh\" (UniqueName: \"kubernetes.io/projected/93cc60c5-709e-474e-81c5-04242bbd0dec-kube-api-access-6bwfh\") pod \"cinder-operator-controller-manager-56ddbfcbfc-bclmv\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.721994 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-apiservice-cert\") pod \"cinder-operator-controller-manager-56ddbfcbfc-bclmv\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.722032 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-webhook-cert\") pod \"cinder-operator-controller-manager-56ddbfcbfc-bclmv\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.729906 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-apiservice-cert\") pod \"cinder-operator-controller-manager-56ddbfcbfc-bclmv\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.729982 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-webhook-cert\") pod \"cinder-operator-controller-manager-56ddbfcbfc-bclmv\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.741568 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bwfh\" (UniqueName: \"kubernetes.io/projected/93cc60c5-709e-474e-81c5-04242bbd0dec-kube-api-access-6bwfh\") pod \"cinder-operator-controller-manager-56ddbfcbfc-bclmv\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:16 crc kubenswrapper[4989]: I1213 17:33:16.812612 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:17 crc kubenswrapper[4989]: I1213 17:33:17.305143 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv"] Dec 13 17:33:17 crc kubenswrapper[4989]: I1213 17:33:17.635280 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" event={"ID":"93cc60c5-709e-474e-81c5-04242bbd0dec","Type":"ContainerStarted","Data":"8467999a1210df50fc886ab6d8a484b810067fab4cbb5b75ed86cfbfebf65fc3"} Dec 13 17:33:21 crc kubenswrapper[4989]: I1213 17:33:21.675701 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" event={"ID":"93cc60c5-709e-474e-81c5-04242bbd0dec","Type":"ContainerStarted","Data":"aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482"} Dec 13 17:33:21 crc kubenswrapper[4989]: I1213 17:33:21.676284 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:21 crc kubenswrapper[4989]: I1213 17:33:21.701245 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" podStartSLOduration=2.041221623 podStartE2EDuration="5.701223574s" podCreationTimestamp="2025-12-13 17:33:16 +0000 UTC" firstStartedPulling="2025-12-13 17:33:17.318205664 +0000 UTC m=+1011.924652802" lastFinishedPulling="2025-12-13 17:33:20.978207615 +0000 UTC m=+1015.584654753" observedRunningTime="2025-12-13 17:33:21.69432942 +0000 UTC m=+1016.300776568" watchObservedRunningTime="2025-12-13 17:33:21.701223574 +0000 UTC m=+1016.307670712" Dec 13 17:33:26 crc kubenswrapper[4989]: I1213 17:33:26.820892 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:33:38 crc kubenswrapper[4989]: I1213 17:33:38.829668 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:33:38 crc kubenswrapper[4989]: I1213 17:33:38.830151 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:33:41 crc kubenswrapper[4989]: I1213 17:33:41.897690 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch"] Dec 13 17:33:41 crc kubenswrapper[4989]: I1213 17:33:41.899142 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" Dec 13 17:33:41 crc kubenswrapper[4989]: I1213 17:33:41.902362 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-db-secret" Dec 13 17:33:41 crc kubenswrapper[4989]: I1213 17:33:41.907808 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-create-ws7zm"] Dec 13 17:33:41 crc kubenswrapper[4989]: I1213 17:33:41.908819 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-ws7zm" Dec 13 17:33:41 crc kubenswrapper[4989]: I1213 17:33:41.912625 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-operator-scripts\") pod \"cinder-2ef4-account-create-update-tp7ch\" (UID: \"5c53a93e-56a6-49e5-9ee4-8113b9bb2693\") " pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" Dec 13 17:33:41 crc kubenswrapper[4989]: I1213 17:33:41.912714 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g55rn\" (UniqueName: \"kubernetes.io/projected/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-kube-api-access-g55rn\") pod \"cinder-2ef4-account-create-update-tp7ch\" (UID: \"5c53a93e-56a6-49e5-9ee4-8113b9bb2693\") " pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" Dec 13 17:33:41 crc kubenswrapper[4989]: I1213 17:33:41.912877 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91505509-42c1-480e-9dbe-384203ec09b5-operator-scripts\") pod \"cinder-db-create-ws7zm\" (UID: \"91505509-42c1-480e-9dbe-384203ec09b5\") " pod="cinder-kuttl-tests/cinder-db-create-ws7zm" Dec 13 17:33:41 crc kubenswrapper[4989]: I1213 17:33:41.912963 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlmbm\" (UniqueName: \"kubernetes.io/projected/91505509-42c1-480e-9dbe-384203ec09b5-kube-api-access-zlmbm\") pod \"cinder-db-create-ws7zm\" (UID: \"91505509-42c1-480e-9dbe-384203ec09b5\") " pod="cinder-kuttl-tests/cinder-db-create-ws7zm" Dec 13 17:33:41 crc kubenswrapper[4989]: I1213 17:33:41.917980 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-ws7zm"] Dec 13 17:33:41 crc kubenswrapper[4989]: I1213 17:33:41.925199 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch"] Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.013596 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlmbm\" (UniqueName: \"kubernetes.io/projected/91505509-42c1-480e-9dbe-384203ec09b5-kube-api-access-zlmbm\") pod \"cinder-db-create-ws7zm\" (UID: \"91505509-42c1-480e-9dbe-384203ec09b5\") " pod="cinder-kuttl-tests/cinder-db-create-ws7zm" Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.013673 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-operator-scripts\") pod \"cinder-2ef4-account-create-update-tp7ch\" (UID: \"5c53a93e-56a6-49e5-9ee4-8113b9bb2693\") " pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.013702 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g55rn\" (UniqueName: \"kubernetes.io/projected/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-kube-api-access-g55rn\") pod \"cinder-2ef4-account-create-update-tp7ch\" (UID: \"5c53a93e-56a6-49e5-9ee4-8113b9bb2693\") " pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.013730 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91505509-42c1-480e-9dbe-384203ec09b5-operator-scripts\") pod \"cinder-db-create-ws7zm\" (UID: \"91505509-42c1-480e-9dbe-384203ec09b5\") " pod="cinder-kuttl-tests/cinder-db-create-ws7zm" Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.014489 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-operator-scripts\") pod \"cinder-2ef4-account-create-update-tp7ch\" (UID: \"5c53a93e-56a6-49e5-9ee4-8113b9bb2693\") " pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.014521 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91505509-42c1-480e-9dbe-384203ec09b5-operator-scripts\") pod \"cinder-db-create-ws7zm\" (UID: \"91505509-42c1-480e-9dbe-384203ec09b5\") " pod="cinder-kuttl-tests/cinder-db-create-ws7zm" Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.033244 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g55rn\" (UniqueName: \"kubernetes.io/projected/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-kube-api-access-g55rn\") pod \"cinder-2ef4-account-create-update-tp7ch\" (UID: \"5c53a93e-56a6-49e5-9ee4-8113b9bb2693\") " pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.033244 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlmbm\" (UniqueName: \"kubernetes.io/projected/91505509-42c1-480e-9dbe-384203ec09b5-kube-api-access-zlmbm\") pod \"cinder-db-create-ws7zm\" (UID: \"91505509-42c1-480e-9dbe-384203ec09b5\") " pod="cinder-kuttl-tests/cinder-db-create-ws7zm" Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.223220 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.240065 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-ws7zm" Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.695613 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-ws7zm"] Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.756492 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch"] Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.834142 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" event={"ID":"5c53a93e-56a6-49e5-9ee4-8113b9bb2693","Type":"ContainerStarted","Data":"3bae409ca94103c3cbe05b33ac88d81aeec9de0994fc76d010d503341377fb7c"} Dec 13 17:33:42 crc kubenswrapper[4989]: I1213 17:33:42.835280 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-ws7zm" event={"ID":"91505509-42c1-480e-9dbe-384203ec09b5","Type":"ContainerStarted","Data":"cb704b4251a280a74da49848971d0d32a9d75a4318aad5b5855978e486092396"} Dec 13 17:33:43 crc kubenswrapper[4989]: I1213 17:33:43.842439 4989 generic.go:334] "Generic (PLEG): container finished" podID="5c53a93e-56a6-49e5-9ee4-8113b9bb2693" containerID="1e5dbc5fdc6f743c02662f552524d0621038235624f8c4c568c58474afdf9afd" exitCode=0 Dec 13 17:33:43 crc kubenswrapper[4989]: I1213 17:33:43.842508 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" event={"ID":"5c53a93e-56a6-49e5-9ee4-8113b9bb2693","Type":"ContainerDied","Data":"1e5dbc5fdc6f743c02662f552524d0621038235624f8c4c568c58474afdf9afd"} Dec 13 17:33:43 crc kubenswrapper[4989]: I1213 17:33:43.844310 4989 generic.go:334] "Generic (PLEG): container finished" podID="91505509-42c1-480e-9dbe-384203ec09b5" containerID="f706b390351e2d289287322d78fbdbcda39572588c81ccc338aee0aa910a6ca9" exitCode=0 Dec 13 17:33:43 crc kubenswrapper[4989]: I1213 17:33:43.844333 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-ws7zm" event={"ID":"91505509-42c1-480e-9dbe-384203ec09b5","Type":"ContainerDied","Data":"f706b390351e2d289287322d78fbdbcda39572588c81ccc338aee0aa910a6ca9"} Dec 13 17:33:44 crc kubenswrapper[4989]: I1213 17:33:44.906971 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.304838 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.305199 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-ws7zm" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.457720 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlmbm\" (UniqueName: \"kubernetes.io/projected/91505509-42c1-480e-9dbe-384203ec09b5-kube-api-access-zlmbm\") pod \"91505509-42c1-480e-9dbe-384203ec09b5\" (UID: \"91505509-42c1-480e-9dbe-384203ec09b5\") " Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.457835 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g55rn\" (UniqueName: \"kubernetes.io/projected/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-kube-api-access-g55rn\") pod \"5c53a93e-56a6-49e5-9ee4-8113b9bb2693\" (UID: \"5c53a93e-56a6-49e5-9ee4-8113b9bb2693\") " Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.457894 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-operator-scripts\") pod \"5c53a93e-56a6-49e5-9ee4-8113b9bb2693\" (UID: \"5c53a93e-56a6-49e5-9ee4-8113b9bb2693\") " Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.457974 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91505509-42c1-480e-9dbe-384203ec09b5-operator-scripts\") pod \"91505509-42c1-480e-9dbe-384203ec09b5\" (UID: \"91505509-42c1-480e-9dbe-384203ec09b5\") " Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.458642 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c53a93e-56a6-49e5-9ee4-8113b9bb2693" (UID: "5c53a93e-56a6-49e5-9ee4-8113b9bb2693"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.458745 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91505509-42c1-480e-9dbe-384203ec09b5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "91505509-42c1-480e-9dbe-384203ec09b5" (UID: "91505509-42c1-480e-9dbe-384203ec09b5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.471032 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-kube-api-access-g55rn" (OuterVolumeSpecName: "kube-api-access-g55rn") pod "5c53a93e-56a6-49e5-9ee4-8113b9bb2693" (UID: "5c53a93e-56a6-49e5-9ee4-8113b9bb2693"). InnerVolumeSpecName "kube-api-access-g55rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.471264 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91505509-42c1-480e-9dbe-384203ec09b5-kube-api-access-zlmbm" (OuterVolumeSpecName: "kube-api-access-zlmbm") pod "91505509-42c1-480e-9dbe-384203ec09b5" (UID: "91505509-42c1-480e-9dbe-384203ec09b5"). InnerVolumeSpecName "kube-api-access-zlmbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.560947 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91505509-42c1-480e-9dbe-384203ec09b5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.561064 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlmbm\" (UniqueName: \"kubernetes.io/projected/91505509-42c1-480e-9dbe-384203ec09b5-kube-api-access-zlmbm\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.561095 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g55rn\" (UniqueName: \"kubernetes.io/projected/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-kube-api-access-g55rn\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.561114 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c53a93e-56a6-49e5-9ee4-8113b9bb2693-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.875802 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-ws7zm" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.875783 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-ws7zm" event={"ID":"91505509-42c1-480e-9dbe-384203ec09b5","Type":"ContainerDied","Data":"cb704b4251a280a74da49848971d0d32a9d75a4318aad5b5855978e486092396"} Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.875934 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb704b4251a280a74da49848971d0d32a9d75a4318aad5b5855978e486092396" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.878447 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" event={"ID":"5c53a93e-56a6-49e5-9ee4-8113b9bb2693","Type":"ContainerDied","Data":"3bae409ca94103c3cbe05b33ac88d81aeec9de0994fc76d010d503341377fb7c"} Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.878500 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bae409ca94103c3cbe05b33ac88d81aeec9de0994fc76d010d503341377fb7c" Dec 13 17:33:45 crc kubenswrapper[4989]: I1213 17:33:45.878569 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.213496 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-2k9tr"] Dec 13 17:33:47 crc kubenswrapper[4989]: E1213 17:33:47.214901 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c53a93e-56a6-49e5-9ee4-8113b9bb2693" containerName="mariadb-account-create-update" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.214984 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c53a93e-56a6-49e5-9ee4-8113b9bb2693" containerName="mariadb-account-create-update" Dec 13 17:33:47 crc kubenswrapper[4989]: E1213 17:33:47.215060 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91505509-42c1-480e-9dbe-384203ec09b5" containerName="mariadb-database-create" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.215116 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91505509-42c1-480e-9dbe-384203ec09b5" containerName="mariadb-database-create" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.215296 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91505509-42c1-480e-9dbe-384203ec09b5" containerName="mariadb-database-create" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.215369 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c53a93e-56a6-49e5-9ee4-8113b9bb2693" containerName="mariadb-account-create-update" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.215880 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.219628 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-qr7rf" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.219964 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.221879 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.227686 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-2k9tr"] Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.402620 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pkzs\" (UniqueName: \"kubernetes.io/projected/54e8ac0e-d364-4244-a054-8f213285bacb-kube-api-access-5pkzs\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.402697 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-db-sync-config-data\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.402766 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54e8ac0e-d364-4244-a054-8f213285bacb-etc-machine-id\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.402840 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-config-data\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.402940 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-scripts\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.504372 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pkzs\" (UniqueName: \"kubernetes.io/projected/54e8ac0e-d364-4244-a054-8f213285bacb-kube-api-access-5pkzs\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.504673 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-db-sync-config-data\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.504856 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54e8ac0e-d364-4244-a054-8f213285bacb-etc-machine-id\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.505014 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-config-data\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.505167 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-scripts\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.505505 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54e8ac0e-d364-4244-a054-8f213285bacb-etc-machine-id\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.510992 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-db-sync-config-data\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.516951 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-config-data\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.517564 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-scripts\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.523003 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pkzs\" (UniqueName: \"kubernetes.io/projected/54e8ac0e-d364-4244-a054-8f213285bacb-kube-api-access-5pkzs\") pod \"cinder-db-sync-2k9tr\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.572554 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.819931 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-2k9tr"] Dec 13 17:33:47 crc kubenswrapper[4989]: I1213 17:33:47.897717 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" event={"ID":"54e8ac0e-d364-4244-a054-8f213285bacb","Type":"ContainerStarted","Data":"87c26eda774ccffb261a32d6afb4b131903d7290b1da85c94eb12b3c0316a187"} Dec 13 17:34:03 crc kubenswrapper[4989]: E1213 17:34:03.383499 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 13 17:34:03 crc kubenswrapper[4989]: E1213 17:34:03.384111 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5pkzs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-2k9tr_cinder-kuttl-tests(54e8ac0e-d364-4244-a054-8f213285bacb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 13 17:34:03 crc kubenswrapper[4989]: E1213 17:34:03.385283 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" podUID="54e8ac0e-d364-4244-a054-8f213285bacb" Dec 13 17:34:04 crc kubenswrapper[4989]: E1213 17:34:04.033032 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" podUID="54e8ac0e-d364-4244-a054-8f213285bacb" Dec 13 17:34:08 crc kubenswrapper[4989]: I1213 17:34:08.829755 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:34:08 crc kubenswrapper[4989]: I1213 17:34:08.830256 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:34:08 crc kubenswrapper[4989]: I1213 17:34:08.830427 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:34:08 crc kubenswrapper[4989]: I1213 17:34:08.831440 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f7a34899e26b30f2d8d01fb7d7814d99ead88a61514d27cbbc8a2e72806061a9"} pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 17:34:08 crc kubenswrapper[4989]: I1213 17:34:08.831546 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" containerID="cri-o://f7a34899e26b30f2d8d01fb7d7814d99ead88a61514d27cbbc8a2e72806061a9" gracePeriod=600 Dec 13 17:34:09 crc kubenswrapper[4989]: I1213 17:34:09.071123 4989 generic.go:334] "Generic (PLEG): container finished" podID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerID="f7a34899e26b30f2d8d01fb7d7814d99ead88a61514d27cbbc8a2e72806061a9" exitCode=0 Dec 13 17:34:09 crc kubenswrapper[4989]: I1213 17:34:09.071168 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerDied","Data":"f7a34899e26b30f2d8d01fb7d7814d99ead88a61514d27cbbc8a2e72806061a9"} Dec 13 17:34:09 crc kubenswrapper[4989]: I1213 17:34:09.071471 4989 scope.go:117] "RemoveContainer" containerID="b6113f2a3ccfb4b50af5284bdcbd31c249a42a87c9815c957aa5451a83308bc9" Dec 13 17:34:10 crc kubenswrapper[4989]: I1213 17:34:10.081541 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerStarted","Data":"7a7aa4339807f600a727da9523920b27ce8d8223ea989c8dc88df4c91d1c4743"} Dec 13 17:34:18 crc kubenswrapper[4989]: I1213 17:34:18.137642 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" event={"ID":"54e8ac0e-d364-4244-a054-8f213285bacb","Type":"ContainerStarted","Data":"47e1793bb65ee7bf294065057b894ba482b4c88971a1740483f994fc1f0f6172"} Dec 13 17:34:18 crc kubenswrapper[4989]: I1213 17:34:18.155574 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" podStartSLOduration=1.466692531 podStartE2EDuration="31.155556953s" podCreationTimestamp="2025-12-13 17:33:47 +0000 UTC" firstStartedPulling="2025-12-13 17:33:47.826874215 +0000 UTC m=+1042.433321353" lastFinishedPulling="2025-12-13 17:34:17.515738647 +0000 UTC m=+1072.122185775" observedRunningTime="2025-12-13 17:34:18.153728937 +0000 UTC m=+1072.760176085" watchObservedRunningTime="2025-12-13 17:34:18.155556953 +0000 UTC m=+1072.762004091" Dec 13 17:34:23 crc kubenswrapper[4989]: I1213 17:34:23.181094 4989 generic.go:334] "Generic (PLEG): container finished" podID="54e8ac0e-d364-4244-a054-8f213285bacb" containerID="47e1793bb65ee7bf294065057b894ba482b4c88971a1740483f994fc1f0f6172" exitCode=0 Dec 13 17:34:23 crc kubenswrapper[4989]: I1213 17:34:23.181268 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" event={"ID":"54e8ac0e-d364-4244-a054-8f213285bacb","Type":"ContainerDied","Data":"47e1793bb65ee7bf294065057b894ba482b4c88971a1740483f994fc1f0f6172"} Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.502607 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.556529 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-scripts\") pod \"54e8ac0e-d364-4244-a054-8f213285bacb\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.556589 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-db-sync-config-data\") pod \"54e8ac0e-d364-4244-a054-8f213285bacb\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.556625 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54e8ac0e-d364-4244-a054-8f213285bacb-etc-machine-id\") pod \"54e8ac0e-d364-4244-a054-8f213285bacb\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.556660 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-config-data\") pod \"54e8ac0e-d364-4244-a054-8f213285bacb\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.556757 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pkzs\" (UniqueName: \"kubernetes.io/projected/54e8ac0e-d364-4244-a054-8f213285bacb-kube-api-access-5pkzs\") pod \"54e8ac0e-d364-4244-a054-8f213285bacb\" (UID: \"54e8ac0e-d364-4244-a054-8f213285bacb\") " Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.556954 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/54e8ac0e-d364-4244-a054-8f213285bacb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "54e8ac0e-d364-4244-a054-8f213285bacb" (UID: "54e8ac0e-d364-4244-a054-8f213285bacb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.562729 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-scripts" (OuterVolumeSpecName: "scripts") pod "54e8ac0e-d364-4244-a054-8f213285bacb" (UID: "54e8ac0e-d364-4244-a054-8f213285bacb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.562996 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "54e8ac0e-d364-4244-a054-8f213285bacb" (UID: "54e8ac0e-d364-4244-a054-8f213285bacb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.563107 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54e8ac0e-d364-4244-a054-8f213285bacb-kube-api-access-5pkzs" (OuterVolumeSpecName: "kube-api-access-5pkzs") pod "54e8ac0e-d364-4244-a054-8f213285bacb" (UID: "54e8ac0e-d364-4244-a054-8f213285bacb"). InnerVolumeSpecName "kube-api-access-5pkzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.596883 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-config-data" (OuterVolumeSpecName: "config-data") pod "54e8ac0e-d364-4244-a054-8f213285bacb" (UID: "54e8ac0e-d364-4244-a054-8f213285bacb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.659053 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.659122 4989 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.659144 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54e8ac0e-d364-4244-a054-8f213285bacb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.659162 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54e8ac0e-d364-4244-a054-8f213285bacb-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:34:24 crc kubenswrapper[4989]: I1213 17:34:24.659178 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pkzs\" (UniqueName: \"kubernetes.io/projected/54e8ac0e-d364-4244-a054-8f213285bacb-kube-api-access-5pkzs\") on node \"crc\" DevicePath \"\"" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.200977 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" event={"ID":"54e8ac0e-d364-4244-a054-8f213285bacb","Type":"ContainerDied","Data":"87c26eda774ccffb261a32d6afb4b131903d7290b1da85c94eb12b3c0316a187"} Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.201499 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87c26eda774ccffb261a32d6afb4b131903d7290b1da85c94eb12b3c0316a187" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.201101 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-2k9tr" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.615217 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:34:25 crc kubenswrapper[4989]: E1213 17:34:25.615679 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54e8ac0e-d364-4244-a054-8f213285bacb" containerName="cinder-db-sync" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.615699 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="54e8ac0e-d364-4244-a054-8f213285bacb" containerName="cinder-db-sync" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.615975 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="54e8ac0e-d364-4244-a054-8f213285bacb" containerName="cinder-db-sync" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.617059 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.620413 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scheduler-config-data" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.620637 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.620669 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.620809 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-qr7rf" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.629981 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.631666 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.636429 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-backup-config-data" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.639210 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.640756 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.642728 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-volume-volume1-config-data" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.645907 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.665708 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.670282 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.671752 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.671826 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.671863 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.671886 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-run\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.671907 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.671929 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdv5t\" (UniqueName: \"kubernetes.io/projected/aea7e175-a80a-4630-9b10-278591652873-kube-api-access-bdv5t\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.671960 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.671995 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672020 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672080 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672101 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-lib-modules\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672124 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672150 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672298 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672546 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgrrb\" (UniqueName: \"kubernetes.io/projected/b4cccfb2-6a7a-456b-a355-1300d7f59fef-kube-api-access-fgrrb\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672602 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672758 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wbcx\" (UniqueName: \"kubernetes.io/projected/0a999f01-db9b-4b04-9b64-1430f1e7eac1-kube-api-access-6wbcx\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672877 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-sys\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672909 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-dev\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672927 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-run\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.672955 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673011 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-scripts\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673041 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673064 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-sys\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673083 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673100 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673123 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673148 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673177 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-dev\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673205 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673223 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673250 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-scripts\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.673271 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b4cccfb2-6a7a-456b-a355-1300d7f59fef-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775074 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775146 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775177 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775198 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775222 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-run\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775247 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdv5t\" (UniqueName: \"kubernetes.io/projected/aea7e175-a80a-4630-9b10-278591652873-kube-api-access-bdv5t\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775290 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775321 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775342 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775360 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775382 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-lib-modules\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775407 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775429 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775449 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775474 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgrrb\" (UniqueName: \"kubernetes.io/projected/b4cccfb2-6a7a-456b-a355-1300d7f59fef-kube-api-access-fgrrb\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775497 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775531 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wbcx\" (UniqueName: \"kubernetes.io/projected/0a999f01-db9b-4b04-9b64-1430f1e7eac1-kube-api-access-6wbcx\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775560 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-sys\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775586 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-dev\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775605 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-run\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775630 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775662 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-scripts\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775690 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775720 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775743 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-sys\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775766 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775811 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775836 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775863 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-dev\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775891 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775915 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775943 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-scripts\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.775967 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b4cccfb2-6a7a-456b-a355-1300d7f59fef-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.776066 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b4cccfb2-6a7a-456b-a355-1300d7f59fef-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.776855 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-sys\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.776905 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-run\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.776936 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-sys\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.776940 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.776917 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-dev\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.776963 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-dev\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777027 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777107 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777123 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777180 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777239 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777204 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777256 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777273 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777490 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777531 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777877 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-lib-modules\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777927 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-run\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777927 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.777972 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.782081 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.782244 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.782721 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-scripts\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.786704 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.787432 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.787632 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.787920 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.788228 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-scripts\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.788834 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.798479 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wbcx\" (UniqueName: \"kubernetes.io/projected/0a999f01-db9b-4b04-9b64-1430f1e7eac1-kube-api-access-6wbcx\") pod \"cinder-backup-0\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.799084 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdv5t\" (UniqueName: \"kubernetes.io/projected/aea7e175-a80a-4630-9b10-278591652873-kube-api-access-bdv5t\") pod \"cinder-volume-volume1-0\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.810176 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgrrb\" (UniqueName: \"kubernetes.io/projected/b4cccfb2-6a7a-456b-a355-1300d7f59fef-kube-api-access-fgrrb\") pod \"cinder-scheduler-0\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.902018 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.903188 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.930306 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.934591 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.939356 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.952522 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.962818 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.978657 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-scripts\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.978721 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f1acff6-4aa7-4860-aa71-298b67012ea4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.978741 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f1acff6-4aa7-4860-aa71-298b67012ea4-logs\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.978766 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwznb\" (UniqueName: \"kubernetes.io/projected/4f1acff6-4aa7-4860-aa71-298b67012ea4-kube-api-access-xwznb\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.978806 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data-custom\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:25 crc kubenswrapper[4989]: I1213 17:34:25.979007 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.080276 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f1acff6-4aa7-4860-aa71-298b67012ea4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.080325 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f1acff6-4aa7-4860-aa71-298b67012ea4-logs\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.080361 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwznb\" (UniqueName: \"kubernetes.io/projected/4f1acff6-4aa7-4860-aa71-298b67012ea4-kube-api-access-xwznb\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.080384 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data-custom\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.080446 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f1acff6-4aa7-4860-aa71-298b67012ea4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.080494 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.080587 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-scripts\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.080847 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f1acff6-4aa7-4860-aa71-298b67012ea4-logs\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.083827 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-scripts\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.091212 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.099119 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.100113 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwznb\" (UniqueName: \"kubernetes.io/projected/4f1acff6-4aa7-4860-aa71-298b67012ea4-kube-api-access-xwznb\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.106263 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data-custom\") pod \"cinder-api-0\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:26 crc kubenswrapper[4989]: I1213 17:34:26.223510 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:28 crc kubenswrapper[4989]: I1213 17:34:28.012537 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:34:28 crc kubenswrapper[4989]: I1213 17:34:28.076030 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:34:28 crc kubenswrapper[4989]: I1213 17:34:28.200982 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:34:28 crc kubenswrapper[4989]: I1213 17:34:28.204998 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:34:28 crc kubenswrapper[4989]: W1213 17:34:28.206235 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f1acff6_4aa7_4860_aa71_298b67012ea4.slice/crio-90f6367f025927cfd303c69f80ca99154671730ead43948b32c55d54435adcff WatchSource:0}: Error finding container 90f6367f025927cfd303c69f80ca99154671730ead43948b32c55d54435adcff: Status 404 returned error can't find the container with id 90f6367f025927cfd303c69f80ca99154671730ead43948b32c55d54435adcff Dec 13 17:34:28 crc kubenswrapper[4989]: W1213 17:34:28.208448 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaea7e175_a80a_4630_9b10_278591652873.slice/crio-5be0815972e3a4cd63f94a9d0df1fb8fb477e7d10dcb528353fb492239737e36 WatchSource:0}: Error finding container 5be0815972e3a4cd63f94a9d0df1fb8fb477e7d10dcb528353fb492239737e36: Status 404 returned error can't find the container with id 5be0815972e3a4cd63f94a9d0df1fb8fb477e7d10dcb528353fb492239737e36 Dec 13 17:34:28 crc kubenswrapper[4989]: I1213 17:34:28.239701 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0a999f01-db9b-4b04-9b64-1430f1e7eac1","Type":"ContainerStarted","Data":"b08d8efbe70d67f14d03dfaacfbd8bdb0c2b1f04b7d884aa184cc930424f0c7d"} Dec 13 17:34:28 crc kubenswrapper[4989]: I1213 17:34:28.240993 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"b4cccfb2-6a7a-456b-a355-1300d7f59fef","Type":"ContainerStarted","Data":"88459dad79e652faee46a01d91b9d33de65c2064940b0a40fae40f9a3db621a1"} Dec 13 17:34:28 crc kubenswrapper[4989]: I1213 17:34:28.241817 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerStarted","Data":"5be0815972e3a4cd63f94a9d0df1fb8fb477e7d10dcb528353fb492239737e36"} Dec 13 17:34:28 crc kubenswrapper[4989]: I1213 17:34:28.242590 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"4f1acff6-4aa7-4860-aa71-298b67012ea4","Type":"ContainerStarted","Data":"90f6367f025927cfd303c69f80ca99154671730ead43948b32c55d54435adcff"} Dec 13 17:34:29 crc kubenswrapper[4989]: I1213 17:34:29.252782 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"4f1acff6-4aa7-4860-aa71-298b67012ea4","Type":"ContainerStarted","Data":"cdcf3191a039b30957514bfec1acf1d206721c732f0d517924316f3ea2088c9b"} Dec 13 17:34:30 crc kubenswrapper[4989]: I1213 17:34:30.262472 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"4f1acff6-4aa7-4860-aa71-298b67012ea4","Type":"ContainerStarted","Data":"411caefce67180e749b5a8c899d65ae03b2209262889a2e7ebd52d3d2831dc29"} Dec 13 17:34:30 crc kubenswrapper[4989]: I1213 17:34:30.263006 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:30 crc kubenswrapper[4989]: I1213 17:34:30.282311 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=5.282230893 podStartE2EDuration="5.282230893s" podCreationTimestamp="2025-12-13 17:34:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:34:30.276835606 +0000 UTC m=+1084.883282744" watchObservedRunningTime="2025-12-13 17:34:30.282230893 +0000 UTC m=+1084.888678031" Dec 13 17:34:31 crc kubenswrapper[4989]: I1213 17:34:31.275488 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0a999f01-db9b-4b04-9b64-1430f1e7eac1","Type":"ContainerStarted","Data":"a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5"} Dec 13 17:34:31 crc kubenswrapper[4989]: I1213 17:34:31.276602 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0a999f01-db9b-4b04-9b64-1430f1e7eac1","Type":"ContainerStarted","Data":"f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56"} Dec 13 17:34:31 crc kubenswrapper[4989]: I1213 17:34:31.279593 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"b4cccfb2-6a7a-456b-a355-1300d7f59fef","Type":"ContainerStarted","Data":"86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224"} Dec 13 17:34:31 crc kubenswrapper[4989]: I1213 17:34:31.285943 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerStarted","Data":"90e95d73aec4371dd624e9a80fe7c66e3a53d3442a538cebae62025801463570"} Dec 13 17:34:31 crc kubenswrapper[4989]: I1213 17:34:31.286010 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerStarted","Data":"56c2ce9da9c132f2e2988130535732a9b2019bffa37177b9b603421f831ba5cd"} Dec 13 17:34:31 crc kubenswrapper[4989]: I1213 17:34:31.314746 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-0" podStartSLOduration=3.726518449 podStartE2EDuration="6.314714996s" podCreationTimestamp="2025-12-13 17:34:25 +0000 UTC" firstStartedPulling="2025-12-13 17:34:28.007215686 +0000 UTC m=+1082.613662824" lastFinishedPulling="2025-12-13 17:34:30.595412233 +0000 UTC m=+1085.201859371" observedRunningTime="2025-12-13 17:34:31.308236845 +0000 UTC m=+1085.914683983" watchObservedRunningTime="2025-12-13 17:34:31.314714996 +0000 UTC m=+1085.921162134" Dec 13 17:34:31 crc kubenswrapper[4989]: I1213 17:34:31.363876 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podStartSLOduration=3.9775933820000002 podStartE2EDuration="6.363839271s" podCreationTimestamp="2025-12-13 17:34:25 +0000 UTC" firstStartedPulling="2025-12-13 17:34:28.210182885 +0000 UTC m=+1082.816630023" lastFinishedPulling="2025-12-13 17:34:30.596428774 +0000 UTC m=+1085.202875912" observedRunningTime="2025-12-13 17:34:31.351943482 +0000 UTC m=+1085.958390620" watchObservedRunningTime="2025-12-13 17:34:31.363839271 +0000 UTC m=+1085.970286409" Dec 13 17:34:32 crc kubenswrapper[4989]: I1213 17:34:32.296748 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"b4cccfb2-6a7a-456b-a355-1300d7f59fef","Type":"ContainerStarted","Data":"e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff"} Dec 13 17:34:32 crc kubenswrapper[4989]: I1213 17:34:32.326771 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-0" podStartSLOduration=4.840287237 podStartE2EDuration="7.326747746s" podCreationTimestamp="2025-12-13 17:34:25 +0000 UTC" firstStartedPulling="2025-12-13 17:34:28.109956885 +0000 UTC m=+1082.716404023" lastFinishedPulling="2025-12-13 17:34:30.596417394 +0000 UTC m=+1085.202864532" observedRunningTime="2025-12-13 17:34:32.321010428 +0000 UTC m=+1086.927457566" watchObservedRunningTime="2025-12-13 17:34:32.326747746 +0000 UTC m=+1086.933194904" Dec 13 17:34:34 crc kubenswrapper[4989]: I1213 17:34:34.316437 4989 generic.go:334] "Generic (PLEG): container finished" podID="aea7e175-a80a-4630-9b10-278591652873" containerID="90e95d73aec4371dd624e9a80fe7c66e3a53d3442a538cebae62025801463570" exitCode=1 Dec 13 17:34:34 crc kubenswrapper[4989]: I1213 17:34:34.316856 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerDied","Data":"90e95d73aec4371dd624e9a80fe7c66e3a53d3442a538cebae62025801463570"} Dec 13 17:34:34 crc kubenswrapper[4989]: I1213 17:34:34.317583 4989 scope.go:117] "RemoveContainer" containerID="90e95d73aec4371dd624e9a80fe7c66e3a53d3442a538cebae62025801463570" Dec 13 17:34:35 crc kubenswrapper[4989]: I1213 17:34:35.327308 4989 generic.go:334] "Generic (PLEG): container finished" podID="aea7e175-a80a-4630-9b10-278591652873" containerID="56c2ce9da9c132f2e2988130535732a9b2019bffa37177b9b603421f831ba5cd" exitCode=1 Dec 13 17:34:35 crc kubenswrapper[4989]: I1213 17:34:35.327448 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerDied","Data":"56c2ce9da9c132f2e2988130535732a9b2019bffa37177b9b603421f831ba5cd"} Dec 13 17:34:35 crc kubenswrapper[4989]: I1213 17:34:35.327816 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerStarted","Data":"09155af02ec6e95d70a3a852c4ce52bd818d90cb1b589c7a04fee1e13a475a2d"} Dec 13 17:34:35 crc kubenswrapper[4989]: I1213 17:34:35.328591 4989 scope.go:117] "RemoveContainer" containerID="56c2ce9da9c132f2e2988130535732a9b2019bffa37177b9b603421f831ba5cd" Dec 13 17:34:35 crc kubenswrapper[4989]: I1213 17:34:35.935778 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:35 crc kubenswrapper[4989]: I1213 17:34:35.953762 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:35 crc kubenswrapper[4989]: I1213 17:34:35.966755 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:35 crc kubenswrapper[4989]: I1213 17:34:35.966829 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:36 crc kubenswrapper[4989]: I1213 17:34:36.271919 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:34:36 crc kubenswrapper[4989]: I1213 17:34:36.355850 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerStarted","Data":"ac916d8accc71ddbbe32ce33b265939f760db944c9c48ff9ad428a52dcd6029d"} Dec 13 17:34:36 crc kubenswrapper[4989]: I1213 17:34:36.356637 4989 scope.go:117] "RemoveContainer" containerID="09155af02ec6e95d70a3a852c4ce52bd818d90cb1b589c7a04fee1e13a475a2d" Dec 13 17:34:36 crc kubenswrapper[4989]: E1213 17:34:36.357446 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:34:37 crc kubenswrapper[4989]: I1213 17:34:37.369309 4989 generic.go:334] "Generic (PLEG): container finished" podID="aea7e175-a80a-4630-9b10-278591652873" containerID="09155af02ec6e95d70a3a852c4ce52bd818d90cb1b589c7a04fee1e13a475a2d" exitCode=1 Dec 13 17:34:37 crc kubenswrapper[4989]: I1213 17:34:37.369366 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerDied","Data":"09155af02ec6e95d70a3a852c4ce52bd818d90cb1b589c7a04fee1e13a475a2d"} Dec 13 17:34:37 crc kubenswrapper[4989]: I1213 17:34:37.369409 4989 scope.go:117] "RemoveContainer" containerID="90e95d73aec4371dd624e9a80fe7c66e3a53d3442a538cebae62025801463570" Dec 13 17:34:37 crc kubenswrapper[4989]: I1213 17:34:37.369994 4989 scope.go:117] "RemoveContainer" containerID="09155af02ec6e95d70a3a852c4ce52bd818d90cb1b589c7a04fee1e13a475a2d" Dec 13 17:34:37 crc kubenswrapper[4989]: E1213 17:34:37.370287 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:34:38 crc kubenswrapper[4989]: I1213 17:34:38.377063 4989 generic.go:334] "Generic (PLEG): container finished" podID="aea7e175-a80a-4630-9b10-278591652873" containerID="ac916d8accc71ddbbe32ce33b265939f760db944c9c48ff9ad428a52dcd6029d" exitCode=1 Dec 13 17:34:38 crc kubenswrapper[4989]: I1213 17:34:38.377411 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerDied","Data":"ac916d8accc71ddbbe32ce33b265939f760db944c9c48ff9ad428a52dcd6029d"} Dec 13 17:34:38 crc kubenswrapper[4989]: I1213 17:34:38.377451 4989 scope.go:117] "RemoveContainer" containerID="56c2ce9da9c132f2e2988130535732a9b2019bffa37177b9b603421f831ba5cd" Dec 13 17:34:38 crc kubenswrapper[4989]: I1213 17:34:38.377719 4989 scope.go:117] "RemoveContainer" containerID="ac916d8accc71ddbbe32ce33b265939f760db944c9c48ff9ad428a52dcd6029d" Dec 13 17:34:38 crc kubenswrapper[4989]: I1213 17:34:38.377752 4989 scope.go:117] "RemoveContainer" containerID="09155af02ec6e95d70a3a852c4ce52bd818d90cb1b589c7a04fee1e13a475a2d" Dec 13 17:34:38 crc kubenswrapper[4989]: E1213 17:34:38.377995 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:34:38 crc kubenswrapper[4989]: I1213 17:34:38.498287 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:34:39 crc kubenswrapper[4989]: I1213 17:34:39.386585 4989 scope.go:117] "RemoveContainer" containerID="ac916d8accc71ddbbe32ce33b265939f760db944c9c48ff9ad428a52dcd6029d" Dec 13 17:34:39 crc kubenswrapper[4989]: I1213 17:34:39.386613 4989 scope.go:117] "RemoveContainer" containerID="09155af02ec6e95d70a3a852c4ce52bd818d90cb1b589c7a04fee1e13a475a2d" Dec 13 17:34:39 crc kubenswrapper[4989]: E1213 17:34:39.386980 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:34:40 crc kubenswrapper[4989]: I1213 17:34:40.963316 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:40 crc kubenswrapper[4989]: I1213 17:34:40.963969 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:40 crc kubenswrapper[4989]: I1213 17:34:40.964297 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:34:40 crc kubenswrapper[4989]: I1213 17:34:40.964827 4989 scope.go:117] "RemoveContainer" containerID="ac916d8accc71ddbbe32ce33b265939f760db944c9c48ff9ad428a52dcd6029d" Dec 13 17:34:40 crc kubenswrapper[4989]: I1213 17:34:40.964858 4989 scope.go:117] "RemoveContainer" containerID="09155af02ec6e95d70a3a852c4ce52bd818d90cb1b589c7a04fee1e13a475a2d" Dec 13 17:34:40 crc kubenswrapper[4989]: E1213 17:34:40.965078 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.190846 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.248898 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.250366 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.260693 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.378445 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.378504 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jk8h\" (UniqueName: \"kubernetes.io/projected/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-kube-api-access-6jk8h\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.378600 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-scripts\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.378665 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.378762 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.404744 4989 scope.go:117] "RemoveContainer" containerID="ac916d8accc71ddbbe32ce33b265939f760db944c9c48ff9ad428a52dcd6029d" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.404769 4989 scope.go:117] "RemoveContainer" containerID="09155af02ec6e95d70a3a852c4ce52bd818d90cb1b589c7a04fee1e13a475a2d" Dec 13 17:34:41 crc kubenswrapper[4989]: E1213 17:34:41.405070 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.480461 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.480520 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.480573 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jk8h\" (UniqueName: \"kubernetes.io/projected/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-kube-api-access-6jk8h\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.480639 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-scripts\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.480742 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.480737 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.486815 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.486868 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.493218 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-scripts\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.495339 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jk8h\" (UniqueName: \"kubernetes.io/projected/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-kube-api-access-6jk8h\") pod \"cinder-scheduler-1\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.575318 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:41 crc kubenswrapper[4989]: I1213 17:34:41.906733 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 13 17:34:41 crc kubenswrapper[4989]: W1213 17:34:41.913831 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ba4bdab_c4c3_4fe2_b4c0_219cb4e0bfe0.slice/crio-818a35ae65452cd2a3c6b0f0d95e2f20829e4ac77bd2c8a9bb238bf0f59e1549 WatchSource:0}: Error finding container 818a35ae65452cd2a3c6b0f0d95e2f20829e4ac77bd2c8a9bb238bf0f59e1549: Status 404 returned error can't find the container with id 818a35ae65452cd2a3c6b0f0d95e2f20829e4ac77bd2c8a9bb238bf0f59e1549 Dec 13 17:34:42 crc kubenswrapper[4989]: I1213 17:34:42.418139 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0","Type":"ContainerStarted","Data":"818a35ae65452cd2a3c6b0f0d95e2f20829e4ac77bd2c8a9bb238bf0f59e1549"} Dec 13 17:34:43 crc kubenswrapper[4989]: I1213 17:34:43.427546 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0","Type":"ContainerStarted","Data":"76b496b4d2010e1a93d599466eee1ba92bbca609ce483a67635902586327696a"} Dec 13 17:34:43 crc kubenswrapper[4989]: I1213 17:34:43.428082 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0","Type":"ContainerStarted","Data":"5d7ab4cff984d052ecfb5a4520ea2ffca44ab56e2f2cf0de233541546c8e8682"} Dec 13 17:34:43 crc kubenswrapper[4989]: I1213 17:34:43.456730 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-1" podStartSLOduration=2.456708021 podStartE2EDuration="2.456708021s" podCreationTimestamp="2025-12-13 17:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:34:43.452053757 +0000 UTC m=+1098.058500885" watchObservedRunningTime="2025-12-13 17:34:43.456708021 +0000 UTC m=+1098.063155179" Dec 13 17:34:46 crc kubenswrapper[4989]: I1213 17:34:46.576531 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:51 crc kubenswrapper[4989]: I1213 17:34:51.803846 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:34:51 crc kubenswrapper[4989]: I1213 17:34:51.878204 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 13 17:34:51 crc kubenswrapper[4989]: I1213 17:34:51.880029 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:51 crc kubenswrapper[4989]: I1213 17:34:51.903839 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 13 17:34:51 crc kubenswrapper[4989]: I1213 17:34:51.959363 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:51 crc kubenswrapper[4989]: I1213 17:34:51.959446 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:51 crc kubenswrapper[4989]: I1213 17:34:51.959494 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-scripts\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:51 crc kubenswrapper[4989]: I1213 17:34:51.959520 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:51 crc kubenswrapper[4989]: I1213 17:34:51.959591 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xqzc\" (UniqueName: \"kubernetes.io/projected/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-kube-api-access-4xqzc\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.060837 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.060917 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-scripts\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.060950 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.060979 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xqzc\" (UniqueName: \"kubernetes.io/projected/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-kube-api-access-4xqzc\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.061045 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.061101 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.066570 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.066666 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-scripts\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.068603 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.080955 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xqzc\" (UniqueName: \"kubernetes.io/projected/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-kube-api-access-4xqzc\") pod \"cinder-scheduler-2\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.204684 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:52 crc kubenswrapper[4989]: I1213 17:34:52.688520 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 13 17:34:53 crc kubenswrapper[4989]: I1213 17:34:53.505962 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b","Type":"ContainerStarted","Data":"c3e25e097d185dc747ca0984edbfb614cb5cb4bbdee06b8c1dfcec0c59e06f90"} Dec 13 17:34:53 crc kubenswrapper[4989]: I1213 17:34:53.506324 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b","Type":"ContainerStarted","Data":"c1c4fcb4001ccabb7f8b908ec655a45cb85753d65e5c75b075c1385c9b6efff1"} Dec 13 17:34:54 crc kubenswrapper[4989]: I1213 17:34:54.517187 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b","Type":"ContainerStarted","Data":"45f81c72e6c07d6254ad9f42a46e4b6db1f38da43a74860af16a70ffb2110fd5"} Dec 13 17:34:54 crc kubenswrapper[4989]: I1213 17:34:54.536152 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-2" podStartSLOduration=3.5361336530000003 podStartE2EDuration="3.536133653s" podCreationTimestamp="2025-12-13 17:34:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:34:54.53147526 +0000 UTC m=+1109.137922398" watchObservedRunningTime="2025-12-13 17:34:54.536133653 +0000 UTC m=+1109.142580791" Dec 13 17:34:56 crc kubenswrapper[4989]: I1213 17:34:56.019086 4989 scope.go:117] "RemoveContainer" containerID="ac916d8accc71ddbbe32ce33b265939f760db944c9c48ff9ad428a52dcd6029d" Dec 13 17:34:56 crc kubenswrapper[4989]: I1213 17:34:56.019121 4989 scope.go:117] "RemoveContainer" containerID="09155af02ec6e95d70a3a852c4ce52bd818d90cb1b589c7a04fee1e13a475a2d" Dec 13 17:34:56 crc kubenswrapper[4989]: I1213 17:34:56.570865 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerStarted","Data":"5a955c224d2c0e8365a661f7c67108f6582640424320b9a01a22fe4200618e04"} Dec 13 17:34:56 crc kubenswrapper[4989]: I1213 17:34:56.571756 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerStarted","Data":"b290fd853b4407b695971b9f4042304167a008062ed3db536bbdaaf2e375b0b9"} Dec 13 17:34:57 crc kubenswrapper[4989]: I1213 17:34:57.205752 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:34:58 crc kubenswrapper[4989]: I1213 17:34:58.599597 4989 generic.go:334] "Generic (PLEG): container finished" podID="aea7e175-a80a-4630-9b10-278591652873" containerID="5a955c224d2c0e8365a661f7c67108f6582640424320b9a01a22fe4200618e04" exitCode=1 Dec 13 17:34:58 crc kubenswrapper[4989]: I1213 17:34:58.599669 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerDied","Data":"5a955c224d2c0e8365a661f7c67108f6582640424320b9a01a22fe4200618e04"} Dec 13 17:34:58 crc kubenswrapper[4989]: I1213 17:34:58.599971 4989 scope.go:117] "RemoveContainer" containerID="09155af02ec6e95d70a3a852c4ce52bd818d90cb1b589c7a04fee1e13a475a2d" Dec 13 17:34:58 crc kubenswrapper[4989]: I1213 17:34:58.600534 4989 scope.go:117] "RemoveContainer" containerID="5a955c224d2c0e8365a661f7c67108f6582640424320b9a01a22fe4200618e04" Dec 13 17:34:58 crc kubenswrapper[4989]: E1213 17:34:58.600862 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:34:59 crc kubenswrapper[4989]: I1213 17:34:59.611584 4989 generic.go:334] "Generic (PLEG): container finished" podID="aea7e175-a80a-4630-9b10-278591652873" containerID="b290fd853b4407b695971b9f4042304167a008062ed3db536bbdaaf2e375b0b9" exitCode=1 Dec 13 17:34:59 crc kubenswrapper[4989]: I1213 17:34:59.611747 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerDied","Data":"b290fd853b4407b695971b9f4042304167a008062ed3db536bbdaaf2e375b0b9"} Dec 13 17:34:59 crc kubenswrapper[4989]: I1213 17:34:59.612280 4989 scope.go:117] "RemoveContainer" containerID="ac916d8accc71ddbbe32ce33b265939f760db944c9c48ff9ad428a52dcd6029d" Dec 13 17:34:59 crc kubenswrapper[4989]: I1213 17:34:59.613190 4989 scope.go:117] "RemoveContainer" containerID="b290fd853b4407b695971b9f4042304167a008062ed3db536bbdaaf2e375b0b9" Dec 13 17:34:59 crc kubenswrapper[4989]: I1213 17:34:59.613271 4989 scope.go:117] "RemoveContainer" containerID="5a955c224d2c0e8365a661f7c67108f6582640424320b9a01a22fe4200618e04" Dec 13 17:34:59 crc kubenswrapper[4989]: E1213 17:34:59.613848 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:35:00 crc kubenswrapper[4989]: I1213 17:35:00.963930 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:35:00 crc kubenswrapper[4989]: I1213 17:35:00.964219 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:35:00 crc kubenswrapper[4989]: I1213 17:35:00.964838 4989 scope.go:117] "RemoveContainer" containerID="b290fd853b4407b695971b9f4042304167a008062ed3db536bbdaaf2e375b0b9" Dec 13 17:35:00 crc kubenswrapper[4989]: I1213 17:35:00.964853 4989 scope.go:117] "RemoveContainer" containerID="5a955c224d2c0e8365a661f7c67108f6582640424320b9a01a22fe4200618e04" Dec 13 17:35:00 crc kubenswrapper[4989]: E1213 17:35:00.966172 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:35:01 crc kubenswrapper[4989]: I1213 17:35:01.963150 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:35:01 crc kubenswrapper[4989]: I1213 17:35:01.963889 4989 scope.go:117] "RemoveContainer" containerID="b290fd853b4407b695971b9f4042304167a008062ed3db536bbdaaf2e375b0b9" Dec 13 17:35:01 crc kubenswrapper[4989]: I1213 17:35:01.963908 4989 scope.go:117] "RemoveContainer" containerID="5a955c224d2c0e8365a661f7c67108f6582640424320b9a01a22fe4200618e04" Dec 13 17:35:01 crc kubenswrapper[4989]: E1213 17:35:01.964155 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:35:02 crc kubenswrapper[4989]: I1213 17:35:02.479113 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:35:03 crc kubenswrapper[4989]: I1213 17:35:03.445665 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 13 17:35:03 crc kubenswrapper[4989]: I1213 17:35:03.446242 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-2" podUID="6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" containerName="cinder-scheduler" containerID="cri-o://c3e25e097d185dc747ca0984edbfb614cb5cb4bbdee06b8c1dfcec0c59e06f90" gracePeriod=30 Dec 13 17:35:03 crc kubenswrapper[4989]: I1213 17:35:03.446279 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-2" podUID="6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" containerName="probe" containerID="cri-o://45f81c72e6c07d6254ad9f42a46e4b6db1f38da43a74860af16a70ffb2110fd5" gracePeriod=30 Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.665546 4989 generic.go:334] "Generic (PLEG): container finished" podID="6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" containerID="45f81c72e6c07d6254ad9f42a46e4b6db1f38da43a74860af16a70ffb2110fd5" exitCode=0 Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.666102 4989 generic.go:334] "Generic (PLEG): container finished" podID="6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" containerID="c3e25e097d185dc747ca0984edbfb614cb5cb4bbdee06b8c1dfcec0c59e06f90" exitCode=0 Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.665624 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b","Type":"ContainerDied","Data":"45f81c72e6c07d6254ad9f42a46e4b6db1f38da43a74860af16a70ffb2110fd5"} Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.666174 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b","Type":"ContainerDied","Data":"c3e25e097d185dc747ca0984edbfb614cb5cb4bbdee06b8c1dfcec0c59e06f90"} Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.749778 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.882756 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-scripts\") pod \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.883591 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xqzc\" (UniqueName: \"kubernetes.io/projected/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-kube-api-access-4xqzc\") pod \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.883662 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data-custom\") pod \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.883859 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data\") pod \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.883891 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-etc-machine-id\") pod \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\" (UID: \"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b\") " Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.884522 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" (UID: "6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.889832 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" (UID: "6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.901307 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-scripts" (OuterVolumeSpecName: "scripts") pod "6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" (UID: "6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.904944 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-kube-api-access-4xqzc" (OuterVolumeSpecName: "kube-api-access-4xqzc") pod "6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" (UID: "6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b"). InnerVolumeSpecName "kube-api-access-4xqzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.969244 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data" (OuterVolumeSpecName: "config-data") pod "6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" (UID: "6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.986267 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.986310 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.986323 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.986333 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xqzc\" (UniqueName: \"kubernetes.io/projected/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-kube-api-access-4xqzc\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:04 crc kubenswrapper[4989]: I1213 17:35:04.986343 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:05 crc kubenswrapper[4989]: I1213 17:35:05.677282 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b","Type":"ContainerDied","Data":"c1c4fcb4001ccabb7f8b908ec655a45cb85753d65e5c75b075c1385c9b6efff1"} Dec 13 17:35:05 crc kubenswrapper[4989]: I1213 17:35:05.677378 4989 scope.go:117] "RemoveContainer" containerID="45f81c72e6c07d6254ad9f42a46e4b6db1f38da43a74860af16a70ffb2110fd5" Dec 13 17:35:05 crc kubenswrapper[4989]: I1213 17:35:05.677393 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 13 17:35:05 crc kubenswrapper[4989]: I1213 17:35:05.701917 4989 scope.go:117] "RemoveContainer" containerID="c3e25e097d185dc747ca0984edbfb614cb5cb4bbdee06b8c1dfcec0c59e06f90" Dec 13 17:35:05 crc kubenswrapper[4989]: I1213 17:35:05.720011 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 13 17:35:05 crc kubenswrapper[4989]: I1213 17:35:05.727648 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 13 17:35:05 crc kubenswrapper[4989]: I1213 17:35:05.734981 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 13 17:35:05 crc kubenswrapper[4989]: I1213 17:35:05.735303 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-1" podUID="4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" containerName="cinder-scheduler" containerID="cri-o://5d7ab4cff984d052ecfb5a4520ea2ffca44ab56e2f2cf0de233541546c8e8682" gracePeriod=30 Dec 13 17:35:05 crc kubenswrapper[4989]: I1213 17:35:05.735506 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-1" podUID="4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" containerName="probe" containerID="cri-o://76b496b4d2010e1a93d599466eee1ba92bbca609ce483a67635902586327696a" gracePeriod=30 Dec 13 17:35:06 crc kubenswrapper[4989]: I1213 17:35:06.023385 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" path="/var/lib/kubelet/pods/6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b/volumes" Dec 13 17:35:06 crc kubenswrapper[4989]: I1213 17:35:06.685754 4989 generic.go:334] "Generic (PLEG): container finished" podID="4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" containerID="76b496b4d2010e1a93d599466eee1ba92bbca609ce483a67635902586327696a" exitCode=0 Dec 13 17:35:06 crc kubenswrapper[4989]: I1213 17:35:06.685842 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0","Type":"ContainerDied","Data":"76b496b4d2010e1a93d599466eee1ba92bbca609ce483a67635902586327696a"} Dec 13 17:35:09 crc kubenswrapper[4989]: I1213 17:35:09.713233 4989 generic.go:334] "Generic (PLEG): container finished" podID="4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" containerID="5d7ab4cff984d052ecfb5a4520ea2ffca44ab56e2f2cf0de233541546c8e8682" exitCode=0 Dec 13 17:35:09 crc kubenswrapper[4989]: I1213 17:35:09.713715 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0","Type":"ContainerDied","Data":"5d7ab4cff984d052ecfb5a4520ea2ffca44ab56e2f2cf0de233541546c8e8682"} Dec 13 17:35:09 crc kubenswrapper[4989]: I1213 17:35:09.937276 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.088531 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-scripts\") pod \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.088611 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data\") pod \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.088681 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data-custom\") pod \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.088743 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jk8h\" (UniqueName: \"kubernetes.io/projected/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-kube-api-access-6jk8h\") pod \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.088865 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-etc-machine-id\") pod \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\" (UID: \"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0\") " Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.089270 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" (UID: "4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.094012 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-kube-api-access-6jk8h" (OuterVolumeSpecName: "kube-api-access-6jk8h") pod "4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" (UID: "4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0"). InnerVolumeSpecName "kube-api-access-6jk8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.094396 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-scripts" (OuterVolumeSpecName: "scripts") pod "4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" (UID: "4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.104051 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" (UID: "4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.155639 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data" (OuterVolumeSpecName: "config-data") pod "4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" (UID: "4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.190680 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.190714 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.190728 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.190741 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jk8h\" (UniqueName: \"kubernetes.io/projected/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-kube-api-access-6jk8h\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.190750 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.721254 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0","Type":"ContainerDied","Data":"818a35ae65452cd2a3c6b0f0d95e2f20829e4ac77bd2c8a9bb238bf0f59e1549"} Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.721322 4989 scope.go:117] "RemoveContainer" containerID="76b496b4d2010e1a93d599466eee1ba92bbca609ce483a67635902586327696a" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.721381 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.763323 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.769801 4989 scope.go:117] "RemoveContainer" containerID="5d7ab4cff984d052ecfb5a4520ea2ffca44ab56e2f2cf0de233541546c8e8682" Dec 13 17:35:10 crc kubenswrapper[4989]: I1213 17:35:10.770826 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 13 17:35:10 crc kubenswrapper[4989]: E1213 17:35:10.806164 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ba4bdab_c4c3_4fe2_b4c0_219cb4e0bfe0.slice\": RecentStats: unable to find data in memory cache]" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.756192 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 13 17:35:11 crc kubenswrapper[4989]: E1213 17:35:11.757512 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" containerName="cinder-scheduler" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.757595 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" containerName="cinder-scheduler" Dec 13 17:35:11 crc kubenswrapper[4989]: E1213 17:35:11.757667 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" containerName="cinder-scheduler" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.757733 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" containerName="cinder-scheduler" Dec 13 17:35:11 crc kubenswrapper[4989]: E1213 17:35:11.757870 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" containerName="probe" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.757937 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" containerName="probe" Dec 13 17:35:11 crc kubenswrapper[4989]: E1213 17:35:11.758024 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" containerName="probe" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.758086 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" containerName="probe" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.758292 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" containerName="probe" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.758364 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" containerName="probe" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.758424 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" containerName="cinder-scheduler" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.758504 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c8b8f0a-ab0f-4da8-b1ba-daf16894b83b" containerName="cinder-scheduler" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.759412 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.779271 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917225 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-sys\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917271 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-nvme\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917297 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-scripts\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917318 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-dev\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917333 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917347 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917368 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917387 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917409 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5mtw\" (UniqueName: \"kubernetes.io/projected/fc158658-d045-4dd7-a144-b2824cde236a-kube-api-access-v5mtw\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917452 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-lib-modules\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917474 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data-custom\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917502 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917521 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:11 crc kubenswrapper[4989]: I1213 17:35:11.917536 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-run\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018275 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018335 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5mtw\" (UniqueName: \"kubernetes.io/projected/fc158658-d045-4dd7-a144-b2824cde236a-kube-api-access-v5mtw\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018400 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018422 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-lib-modules\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018465 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data-custom\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018511 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018551 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018604 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-run\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018657 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-sys\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018698 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-nvme\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018737 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-scripts\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018783 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018781 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-dev\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018873 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018894 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018908 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018952 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018965 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-nvme\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.018517 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-lib-modules\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.022367 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-run\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.022408 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.022439 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-sys\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.022648 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.025235 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-dev\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.025405 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data-custom\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.027833 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-scripts\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.047625 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0" path="/var/lib/kubelet/pods/4ba4bdab-c4c3-4fe2-b4c0-219cb4e0bfe0/volumes" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.050280 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.052057 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5mtw\" (UniqueName: \"kubernetes.io/projected/fc158658-d045-4dd7-a144-b2824cde236a-kube-api-access-v5mtw\") pod \"cinder-backup-1\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.080746 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.319035 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.737958 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"fc158658-d045-4dd7-a144-b2824cde236a","Type":"ContainerStarted","Data":"796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484"} Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.738417 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"fc158658-d045-4dd7-a144-b2824cde236a","Type":"ContainerStarted","Data":"a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7"} Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.738456 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"fc158658-d045-4dd7-a144-b2824cde236a","Type":"ContainerStarted","Data":"9a60814efaeea77e211ce57771879384674451c37c0259c1907c5c48c1234078"} Dec 13 17:35:12 crc kubenswrapper[4989]: I1213 17:35:12.762631 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-1" podStartSLOduration=1.762607922 podStartE2EDuration="1.762607922s" podCreationTimestamp="2025-12-13 17:35:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:35:12.760231179 +0000 UTC m=+1127.366678337" watchObservedRunningTime="2025-12-13 17:35:12.762607922 +0000 UTC m=+1127.369055070" Dec 13 17:35:15 crc kubenswrapper[4989]: I1213 17:35:15.014577 4989 scope.go:117] "RemoveContainer" containerID="b290fd853b4407b695971b9f4042304167a008062ed3db536bbdaaf2e375b0b9" Dec 13 17:35:15 crc kubenswrapper[4989]: I1213 17:35:15.015081 4989 scope.go:117] "RemoveContainer" containerID="5a955c224d2c0e8365a661f7c67108f6582640424320b9a01a22fe4200618e04" Dec 13 17:35:15 crc kubenswrapper[4989]: E1213 17:35:15.015286 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.082526 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.309278 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.825051 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.826452 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.837184 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914027 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59lsl\" (UniqueName: \"kubernetes.io/projected/1e74c732-e15c-4ed0-81c7-d29cc897ef80-kube-api-access-59lsl\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914158 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914200 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-run\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914249 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-dev\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914311 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914358 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914462 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-sys\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914487 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914550 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-scripts\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914581 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-lib-modules\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914864 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data-custom\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.914927 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.915019 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:17 crc kubenswrapper[4989]: I1213 17:35:17.915063 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-nvme\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017050 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data-custom\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017112 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017160 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017173 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017190 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-nvme\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017216 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59lsl\" (UniqueName: \"kubernetes.io/projected/1e74c732-e15c-4ed0-81c7-d29cc897ef80-kube-api-access-59lsl\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017268 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017325 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-run\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017339 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-nvme\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017353 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-dev\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017380 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017404 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017420 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-run\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017470 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017390 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-dev\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017472 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017439 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017536 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-sys\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017560 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017592 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-scripts\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017617 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017633 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-sys\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017636 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-lib-modules\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.017662 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-lib-modules\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.023429 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-scripts\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.024978 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.025186 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data-custom\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.036480 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59lsl\" (UniqueName: \"kubernetes.io/projected/1e74c732-e15c-4ed0-81c7-d29cc897ef80-kube-api-access-59lsl\") pod \"cinder-backup-2\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.143105 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.409626 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.782692 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"1e74c732-e15c-4ed0-81c7-d29cc897ef80","Type":"ContainerStarted","Data":"0298e2e994cb5e664adc5161932fde7c4cbc8ab17b578817e989ba9fda45ea79"} Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.783126 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"1e74c732-e15c-4ed0-81c7-d29cc897ef80","Type":"ContainerStarted","Data":"7cf539f3738ce1850c43448d5bec21799ae67d9c6b721ead0a9718075a257b53"} Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.783138 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"1e74c732-e15c-4ed0-81c7-d29cc897ef80","Type":"ContainerStarted","Data":"183452ae2bee95a9aa49eb23dba086ec85c61a88d370825c8dd26a4552682750"} Dec 13 17:35:18 crc kubenswrapper[4989]: I1213 17:35:18.806630 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-2" podStartSLOduration=1.8066144240000002 podStartE2EDuration="1.806614424s" podCreationTimestamp="2025-12-13 17:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:35:18.805676155 +0000 UTC m=+1133.412123313" watchObservedRunningTime="2025-12-13 17:35:18.806614424 +0000 UTC m=+1133.413061562" Dec 13 17:35:23 crc kubenswrapper[4989]: I1213 17:35:23.144239 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:23 crc kubenswrapper[4989]: I1213 17:35:23.392146 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:24 crc kubenswrapper[4989]: I1213 17:35:24.147957 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 13 17:35:24 crc kubenswrapper[4989]: I1213 17:35:24.831541 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-2" podUID="1e74c732-e15c-4ed0-81c7-d29cc897ef80" containerName="cinder-backup" containerID="cri-o://7cf539f3738ce1850c43448d5bec21799ae67d9c6b721ead0a9718075a257b53" gracePeriod=30 Dec 13 17:35:24 crc kubenswrapper[4989]: I1213 17:35:24.831604 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-2" podUID="1e74c732-e15c-4ed0-81c7-d29cc897ef80" containerName="probe" containerID="cri-o://0298e2e994cb5e664adc5161932fde7c4cbc8ab17b578817e989ba9fda45ea79" gracePeriod=30 Dec 13 17:35:25 crc kubenswrapper[4989]: I1213 17:35:25.845271 4989 generic.go:334] "Generic (PLEG): container finished" podID="1e74c732-e15c-4ed0-81c7-d29cc897ef80" containerID="0298e2e994cb5e664adc5161932fde7c4cbc8ab17b578817e989ba9fda45ea79" exitCode=0 Dec 13 17:35:25 crc kubenswrapper[4989]: I1213 17:35:25.845319 4989 generic.go:334] "Generic (PLEG): container finished" podID="1e74c732-e15c-4ed0-81c7-d29cc897ef80" containerID="7cf539f3738ce1850c43448d5bec21799ae67d9c6b721ead0a9718075a257b53" exitCode=0 Dec 13 17:35:25 crc kubenswrapper[4989]: I1213 17:35:25.845351 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"1e74c732-e15c-4ed0-81c7-d29cc897ef80","Type":"ContainerDied","Data":"0298e2e994cb5e664adc5161932fde7c4cbc8ab17b578817e989ba9fda45ea79"} Dec 13 17:35:25 crc kubenswrapper[4989]: I1213 17:35:25.845415 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"1e74c732-e15c-4ed0-81c7-d29cc897ef80","Type":"ContainerDied","Data":"7cf539f3738ce1850c43448d5bec21799ae67d9c6b721ead0a9718075a257b53"} Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.026434 4989 scope.go:117] "RemoveContainer" containerID="b290fd853b4407b695971b9f4042304167a008062ed3db536bbdaaf2e375b0b9" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.026897 4989 scope.go:117] "RemoveContainer" containerID="5a955c224d2c0e8365a661f7c67108f6582640424320b9a01a22fe4200618e04" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.100568 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.261400 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-nvme\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262076 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-iscsi\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.261819 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262173 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-cinder\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262250 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-dev\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262319 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data-custom\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262331 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262346 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-machine-id\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262359 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-dev" (OuterVolumeSpecName: "dev") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262390 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-brick\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262429 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59lsl\" (UniqueName: \"kubernetes.io/projected/1e74c732-e15c-4ed0-81c7-d29cc897ef80-kube-api-access-59lsl\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262499 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-scripts\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262448 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262491 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262880 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-sys\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262909 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-lib-cinder\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.262931 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-run\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.263001 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-lib-modules\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.263084 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data\") pod \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\" (UID: \"1e74c732-e15c-4ed0-81c7-d29cc897ef80\") " Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.263091 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.263389 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-sys" (OuterVolumeSpecName: "sys") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.263424 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-run" (OuterVolumeSpecName: "run") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.263571 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.263918 4989 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.263964 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.263980 4989 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-dev\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.263992 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.264004 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.264117 4989 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-sys\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.264129 4989 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.264138 4989 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-run\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.264146 4989 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.267227 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.270283 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e74c732-e15c-4ed0-81c7-d29cc897ef80-kube-api-access-59lsl" (OuterVolumeSpecName: "kube-api-access-59lsl") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "kube-api-access-59lsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.270493 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.274902 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-scripts" (OuterVolumeSpecName: "scripts") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.349075 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data" (OuterVolumeSpecName: "config-data") pod "1e74c732-e15c-4ed0-81c7-d29cc897ef80" (UID: "1e74c732-e15c-4ed0-81c7-d29cc897ef80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.365926 4989 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1e74c732-e15c-4ed0-81c7-d29cc897ef80-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.365975 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.365993 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59lsl\" (UniqueName: \"kubernetes.io/projected/1e74c732-e15c-4ed0-81c7-d29cc897ef80-kube-api-access-59lsl\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.366007 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.366019 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e74c732-e15c-4ed0-81c7-d29cc897ef80-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.862781 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerStarted","Data":"a81daec09666018505c8871efbeb99f8f40e71e3acd5ecb94c2466d22bbb8dd8"} Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.863589 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerStarted","Data":"3452a91b024721df6d1e71e9758645f089c80d3938d3dc800591f31be452eeda"} Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.866779 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"1e74c732-e15c-4ed0-81c7-d29cc897ef80","Type":"ContainerDied","Data":"183452ae2bee95a9aa49eb23dba086ec85c61a88d370825c8dd26a4552682750"} Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.866874 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.866934 4989 scope.go:117] "RemoveContainer" containerID="0298e2e994cb5e664adc5161932fde7c4cbc8ab17b578817e989ba9fda45ea79" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.897429 4989 scope.go:117] "RemoveContainer" containerID="7cf539f3738ce1850c43448d5bec21799ae67d9c6b721ead0a9718075a257b53" Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.917096 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.927470 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.932884 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.933337 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-1" podUID="fc158658-d045-4dd7-a144-b2824cde236a" containerName="cinder-backup" containerID="cri-o://a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7" gracePeriod=30 Dec 13 17:35:26 crc kubenswrapper[4989]: I1213 17:35:26.933410 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-1" podUID="fc158658-d045-4dd7-a144-b2824cde236a" containerName="probe" containerID="cri-o://796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484" gracePeriod=30 Dec 13 17:35:27 crc kubenswrapper[4989]: I1213 17:35:27.876508 4989 generic.go:334] "Generic (PLEG): container finished" podID="fc158658-d045-4dd7-a144-b2824cde236a" containerID="796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484" exitCode=0 Dec 13 17:35:27 crc kubenswrapper[4989]: I1213 17:35:27.876946 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"fc158658-d045-4dd7-a144-b2824cde236a","Type":"ContainerDied","Data":"796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484"} Dec 13 17:35:28 crc kubenswrapper[4989]: I1213 17:35:28.021839 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e74c732-e15c-4ed0-81c7-d29cc897ef80" path="/var/lib/kubelet/pods/1e74c732-e15c-4ed0-81c7-d29cc897ef80/volumes" Dec 13 17:35:28 crc kubenswrapper[4989]: I1213 17:35:28.886041 4989 generic.go:334] "Generic (PLEG): container finished" podID="aea7e175-a80a-4630-9b10-278591652873" containerID="a81daec09666018505c8871efbeb99f8f40e71e3acd5ecb94c2466d22bbb8dd8" exitCode=1 Dec 13 17:35:28 crc kubenswrapper[4989]: I1213 17:35:28.886308 4989 generic.go:334] "Generic (PLEG): container finished" podID="aea7e175-a80a-4630-9b10-278591652873" containerID="3452a91b024721df6d1e71e9758645f089c80d3938d3dc800591f31be452eeda" exitCode=1 Dec 13 17:35:28 crc kubenswrapper[4989]: I1213 17:35:28.886125 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerDied","Data":"a81daec09666018505c8871efbeb99f8f40e71e3acd5ecb94c2466d22bbb8dd8"} Dec 13 17:35:28 crc kubenswrapper[4989]: I1213 17:35:28.886346 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerDied","Data":"3452a91b024721df6d1e71e9758645f089c80d3938d3dc800591f31be452eeda"} Dec 13 17:35:28 crc kubenswrapper[4989]: I1213 17:35:28.886368 4989 scope.go:117] "RemoveContainer" containerID="5a955c224d2c0e8365a661f7c67108f6582640424320b9a01a22fe4200618e04" Dec 13 17:35:28 crc kubenswrapper[4989]: I1213 17:35:28.886971 4989 scope.go:117] "RemoveContainer" containerID="3452a91b024721df6d1e71e9758645f089c80d3938d3dc800591f31be452eeda" Dec 13 17:35:28 crc kubenswrapper[4989]: I1213 17:35:28.887018 4989 scope.go:117] "RemoveContainer" containerID="a81daec09666018505c8871efbeb99f8f40e71e3acd5ecb94c2466d22bbb8dd8" Dec 13 17:35:28 crc kubenswrapper[4989]: E1213 17:35:28.887275 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:35:28 crc kubenswrapper[4989]: I1213 17:35:28.938256 4989 scope.go:117] "RemoveContainer" containerID="b290fd853b4407b695971b9f4042304167a008062ed3db536bbdaaf2e375b0b9" Dec 13 17:35:28 crc kubenswrapper[4989]: I1213 17:35:28.963418 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.640304 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733480 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5mtw\" (UniqueName: \"kubernetes.io/projected/fc158658-d045-4dd7-a144-b2824cde236a-kube-api-access-v5mtw\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733541 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-brick\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733571 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-scripts\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733622 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-iscsi\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733677 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-lib-modules\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733695 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-lib-cinder\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733671 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733694 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733773 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733812 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-run\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733801 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733835 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-dev\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733818 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-run" (OuterVolumeSpecName: "run") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733864 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-cinder\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733881 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-machine-id\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733910 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-dev" (OuterVolumeSpecName: "dev") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733914 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-nvme\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733930 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733947 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733977 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data-custom\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.734002 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-sys\") pod \"fc158658-d045-4dd7-a144-b2824cde236a\" (UID: \"fc158658-d045-4dd7-a144-b2824cde236a\") " Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733941 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.733951 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.734249 4989 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.734259 4989 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.734269 4989 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-run\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.734277 4989 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-dev\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.734284 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.734292 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.734300 4989 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.734308 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.734316 4989 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.734337 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-sys" (OuterVolumeSpecName: "sys") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.740275 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-scripts" (OuterVolumeSpecName: "scripts") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.740693 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc158658-d045-4dd7-a144-b2824cde236a-kube-api-access-v5mtw" (OuterVolumeSpecName: "kube-api-access-v5mtw") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "kube-api-access-v5mtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.741300 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.804917 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data" (OuterVolumeSpecName: "config-data") pod "fc158658-d045-4dd7-a144-b2824cde236a" (UID: "fc158658-d045-4dd7-a144-b2824cde236a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.836717 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.836769 4989 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fc158658-d045-4dd7-a144-b2824cde236a-sys\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.836802 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5mtw\" (UniqueName: \"kubernetes.io/projected/fc158658-d045-4dd7-a144-b2824cde236a-kube-api-access-v5mtw\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.836824 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.836837 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc158658-d045-4dd7-a144-b2824cde236a-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.900681 4989 scope.go:117] "RemoveContainer" containerID="3452a91b024721df6d1e71e9758645f089c80d3938d3dc800591f31be452eeda" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.902674 4989 scope.go:117] "RemoveContainer" containerID="a81daec09666018505c8871efbeb99f8f40e71e3acd5ecb94c2466d22bbb8dd8" Dec 13 17:35:29 crc kubenswrapper[4989]: E1213 17:35:29.903849 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.921521 4989 generic.go:334] "Generic (PLEG): container finished" podID="fc158658-d045-4dd7-a144-b2824cde236a" containerID="a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7" exitCode=0 Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.921580 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"fc158658-d045-4dd7-a144-b2824cde236a","Type":"ContainerDied","Data":"a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7"} Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.921617 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"fc158658-d045-4dd7-a144-b2824cde236a","Type":"ContainerDied","Data":"9a60814efaeea77e211ce57771879384674451c37c0259c1907c5c48c1234078"} Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.921642 4989 scope.go:117] "RemoveContainer" containerID="796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.921690 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.953468 4989 scope.go:117] "RemoveContainer" containerID="a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.966195 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.971841 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.981424 4989 scope.go:117] "RemoveContainer" containerID="796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484" Dec 13 17:35:29 crc kubenswrapper[4989]: E1213 17:35:29.981823 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484\": container with ID starting with 796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484 not found: ID does not exist" containerID="796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.981878 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484"} err="failed to get container status \"796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484\": rpc error: code = NotFound desc = could not find container \"796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484\": container with ID starting with 796f8d69986a208778f2e0420f33e4c0e571e8156f8f1e171606257329d7e484 not found: ID does not exist" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.981910 4989 scope.go:117] "RemoveContainer" containerID="a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7" Dec 13 17:35:29 crc kubenswrapper[4989]: E1213 17:35:29.982342 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7\": container with ID starting with a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7 not found: ID does not exist" containerID="a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7" Dec 13 17:35:29 crc kubenswrapper[4989]: I1213 17:35:29.982370 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7"} err="failed to get container status \"a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7\": rpc error: code = NotFound desc = could not find container \"a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7\": container with ID starting with a6ac77e4c68e4b631b9d0dba98007dfbfdb07c90edd2b1b80ef4d23636cc56f7 not found: ID does not exist" Dec 13 17:35:30 crc kubenswrapper[4989]: I1213 17:35:30.023210 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc158658-d045-4dd7-a144-b2824cde236a" path="/var/lib/kubelet/pods/fc158658-d045-4dd7-a144-b2824cde236a/volumes" Dec 13 17:35:30 crc kubenswrapper[4989]: I1213 17:35:30.414826 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:35:30 crc kubenswrapper[4989]: I1213 17:35:30.415161 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="4f1acff6-4aa7-4860-aa71-298b67012ea4" containerName="cinder-api-log" containerID="cri-o://cdcf3191a039b30957514bfec1acf1d206721c732f0d517924316f3ea2088c9b" gracePeriod=30 Dec 13 17:35:30 crc kubenswrapper[4989]: I1213 17:35:30.415277 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="4f1acff6-4aa7-4860-aa71-298b67012ea4" containerName="cinder-api" containerID="cri-o://411caefce67180e749b5a8c899d65ae03b2209262889a2e7ebd52d3d2831dc29" gracePeriod=30 Dec 13 17:35:30 crc kubenswrapper[4989]: I1213 17:35:30.936060 4989 generic.go:334] "Generic (PLEG): container finished" podID="4f1acff6-4aa7-4860-aa71-298b67012ea4" containerID="cdcf3191a039b30957514bfec1acf1d206721c732f0d517924316f3ea2088c9b" exitCode=143 Dec 13 17:35:30 crc kubenswrapper[4989]: I1213 17:35:30.936113 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"4f1acff6-4aa7-4860-aa71-298b67012ea4","Type":"ContainerDied","Data":"cdcf3191a039b30957514bfec1acf1d206721c732f0d517924316f3ea2088c9b"} Dec 13 17:35:30 crc kubenswrapper[4989]: I1213 17:35:30.962995 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:35:30 crc kubenswrapper[4989]: I1213 17:35:30.963114 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:35:30 crc kubenswrapper[4989]: I1213 17:35:30.964328 4989 scope.go:117] "RemoveContainer" containerID="3452a91b024721df6d1e71e9758645f089c80d3938d3dc800591f31be452eeda" Dec 13 17:35:30 crc kubenswrapper[4989]: I1213 17:35:30.964371 4989 scope.go:117] "RemoveContainer" containerID="a81daec09666018505c8871efbeb99f8f40e71e3acd5ecb94c2466d22bbb8dd8" Dec 13 17:35:30 crc kubenswrapper[4989]: E1213 17:35:30.964918 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:35:33 crc kubenswrapper[4989]: I1213 17:35:33.556436 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="4f1acff6-4aa7-4860-aa71-298b67012ea4" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.93:8776/healthcheck\": read tcp 10.217.0.2:42320->10.217.0.93:8776: read: connection reset by peer" Dec 13 17:35:33 crc kubenswrapper[4989]: I1213 17:35:33.962857 4989 generic.go:334] "Generic (PLEG): container finished" podID="4f1acff6-4aa7-4860-aa71-298b67012ea4" containerID="411caefce67180e749b5a8c899d65ae03b2209262889a2e7ebd52d3d2831dc29" exitCode=0 Dec 13 17:35:33 crc kubenswrapper[4989]: I1213 17:35:33.962938 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"4f1acff6-4aa7-4860-aa71-298b67012ea4","Type":"ContainerDied","Data":"411caefce67180e749b5a8c899d65ae03b2209262889a2e7ebd52d3d2831dc29"} Dec 13 17:35:33 crc kubenswrapper[4989]: I1213 17:35:33.963247 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"4f1acff6-4aa7-4860-aa71-298b67012ea4","Type":"ContainerDied","Data":"90f6367f025927cfd303c69f80ca99154671730ead43948b32c55d54435adcff"} Dec 13 17:35:33 crc kubenswrapper[4989]: I1213 17:35:33.963267 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90f6367f025927cfd303c69f80ca99154671730ead43948b32c55d54435adcff" Dec 13 17:35:33 crc kubenswrapper[4989]: I1213 17:35:33.978623 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.103522 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwznb\" (UniqueName: \"kubernetes.io/projected/4f1acff6-4aa7-4860-aa71-298b67012ea4-kube-api-access-xwznb\") pod \"4f1acff6-4aa7-4860-aa71-298b67012ea4\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.103718 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data-custom\") pod \"4f1acff6-4aa7-4860-aa71-298b67012ea4\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.103783 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f1acff6-4aa7-4860-aa71-298b67012ea4-etc-machine-id\") pod \"4f1acff6-4aa7-4860-aa71-298b67012ea4\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.103898 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data\") pod \"4f1acff6-4aa7-4860-aa71-298b67012ea4\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.104075 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1acff6-4aa7-4860-aa71-298b67012ea4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4f1acff6-4aa7-4860-aa71-298b67012ea4" (UID: "4f1acff6-4aa7-4860-aa71-298b67012ea4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.104200 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-scripts\") pod \"4f1acff6-4aa7-4860-aa71-298b67012ea4\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.104286 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f1acff6-4aa7-4860-aa71-298b67012ea4-logs\") pod \"4f1acff6-4aa7-4860-aa71-298b67012ea4\" (UID: \"4f1acff6-4aa7-4860-aa71-298b67012ea4\") " Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.104753 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f1acff6-4aa7-4860-aa71-298b67012ea4-logs" (OuterVolumeSpecName: "logs") pod "4f1acff6-4aa7-4860-aa71-298b67012ea4" (UID: "4f1acff6-4aa7-4860-aa71-298b67012ea4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.107611 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f1acff6-4aa7-4860-aa71-298b67012ea4-logs\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.107662 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4f1acff6-4aa7-4860-aa71-298b67012ea4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.114939 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4f1acff6-4aa7-4860-aa71-298b67012ea4" (UID: "4f1acff6-4aa7-4860-aa71-298b67012ea4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.114994 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-scripts" (OuterVolumeSpecName: "scripts") pod "4f1acff6-4aa7-4860-aa71-298b67012ea4" (UID: "4f1acff6-4aa7-4860-aa71-298b67012ea4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.116132 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f1acff6-4aa7-4860-aa71-298b67012ea4-kube-api-access-xwznb" (OuterVolumeSpecName: "kube-api-access-xwznb") pod "4f1acff6-4aa7-4860-aa71-298b67012ea4" (UID: "4f1acff6-4aa7-4860-aa71-298b67012ea4"). InnerVolumeSpecName "kube-api-access-xwznb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.152764 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data" (OuterVolumeSpecName: "config-data") pod "4f1acff6-4aa7-4860-aa71-298b67012ea4" (UID: "4f1acff6-4aa7-4860-aa71-298b67012ea4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.208880 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwznb\" (UniqueName: \"kubernetes.io/projected/4f1acff6-4aa7-4860-aa71-298b67012ea4-kube-api-access-xwznb\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.208927 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.208942 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.208952 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f1acff6-4aa7-4860-aa71-298b67012ea4-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:34 crc kubenswrapper[4989]: I1213 17:35:34.970856 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.016764 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.027341 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.680368 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:35:35 crc kubenswrapper[4989]: E1213 17:35:35.681393 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc158658-d045-4dd7-a144-b2824cde236a" containerName="probe" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681427 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc158658-d045-4dd7-a144-b2824cde236a" containerName="probe" Dec 13 17:35:35 crc kubenswrapper[4989]: E1213 17:35:35.681452 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1acff6-4aa7-4860-aa71-298b67012ea4" containerName="cinder-api-log" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681467 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1acff6-4aa7-4860-aa71-298b67012ea4" containerName="cinder-api-log" Dec 13 17:35:35 crc kubenswrapper[4989]: E1213 17:35:35.681490 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e74c732-e15c-4ed0-81c7-d29cc897ef80" containerName="probe" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681504 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e74c732-e15c-4ed0-81c7-d29cc897ef80" containerName="probe" Dec 13 17:35:35 crc kubenswrapper[4989]: E1213 17:35:35.681526 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1acff6-4aa7-4860-aa71-298b67012ea4" containerName="cinder-api" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681538 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1acff6-4aa7-4860-aa71-298b67012ea4" containerName="cinder-api" Dec 13 17:35:35 crc kubenswrapper[4989]: E1213 17:35:35.681567 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e74c732-e15c-4ed0-81c7-d29cc897ef80" containerName="cinder-backup" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681580 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e74c732-e15c-4ed0-81c7-d29cc897ef80" containerName="cinder-backup" Dec 13 17:35:35 crc kubenswrapper[4989]: E1213 17:35:35.681610 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc158658-d045-4dd7-a144-b2824cde236a" containerName="cinder-backup" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681624 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc158658-d045-4dd7-a144-b2824cde236a" containerName="cinder-backup" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681856 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e74c732-e15c-4ed0-81c7-d29cc897ef80" containerName="probe" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681878 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc158658-d045-4dd7-a144-b2824cde236a" containerName="cinder-backup" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681891 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e74c732-e15c-4ed0-81c7-d29cc897ef80" containerName="cinder-backup" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681927 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1acff6-4aa7-4860-aa71-298b67012ea4" containerName="cinder-api" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681947 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1acff6-4aa7-4860-aa71-298b67012ea4" containerName="cinder-api-log" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.681963 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc158658-d045-4dd7-a144-b2824cde236a" containerName="probe" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.684733 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.689915 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.696655 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.704107 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.707376 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.712181 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.714161 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.718231 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.768440 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.835807 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdd1b143-6da9-463d-92a7-ff4c054e89fd-logs\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.835932 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data-custom\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.835967 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.835994 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v7h8\" (UniqueName: \"kubernetes.io/projected/cdd1b143-6da9-463d-92a7-ff4c054e89fd-kube-api-access-6v7h8\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836021 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data-custom\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836053 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16c9922f-a4a5-445f-a2b7-f252d25d1815-etc-machine-id\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836084 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cdd1b143-6da9-463d-92a7-ff4c054e89fd-etc-machine-id\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836240 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data-custom\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836292 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0532be72-ce7d-49fb-a8e7-ed43970792bd-etc-machine-id\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836315 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0532be72-ce7d-49fb-a8e7-ed43970792bd-logs\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836414 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c9922f-a4a5-445f-a2b7-f252d25d1815-logs\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836490 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq6f4\" (UniqueName: \"kubernetes.io/projected/16c9922f-a4a5-445f-a2b7-f252d25d1815-kube-api-access-bq6f4\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836520 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836558 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5thx\" (UniqueName: \"kubernetes.io/projected/0532be72-ce7d-49fb-a8e7-ed43970792bd-kube-api-access-z5thx\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836593 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-scripts\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836663 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-scripts\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836709 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.836774 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-scripts\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938593 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data-custom\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938651 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0532be72-ce7d-49fb-a8e7-ed43970792bd-etc-machine-id\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938675 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0532be72-ce7d-49fb-a8e7-ed43970792bd-logs\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938715 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c9922f-a4a5-445f-a2b7-f252d25d1815-logs\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938754 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq6f4\" (UniqueName: \"kubernetes.io/projected/16c9922f-a4a5-445f-a2b7-f252d25d1815-kube-api-access-bq6f4\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938780 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938836 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5thx\" (UniqueName: \"kubernetes.io/projected/0532be72-ce7d-49fb-a8e7-ed43970792bd-kube-api-access-z5thx\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938859 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-scripts\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938890 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-scripts\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938915 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938948 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-scripts\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938981 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdd1b143-6da9-463d-92a7-ff4c054e89fd-logs\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.938984 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0532be72-ce7d-49fb-a8e7-ed43970792bd-etc-machine-id\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.939043 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data-custom\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.939195 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.939258 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v7h8\" (UniqueName: \"kubernetes.io/projected/cdd1b143-6da9-463d-92a7-ff4c054e89fd-kube-api-access-6v7h8\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.939316 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data-custom\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.939391 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16c9922f-a4a5-445f-a2b7-f252d25d1815-etc-machine-id\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.939480 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cdd1b143-6da9-463d-92a7-ff4c054e89fd-etc-machine-id\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.939857 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cdd1b143-6da9-463d-92a7-ff4c054e89fd-etc-machine-id\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.940260 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0532be72-ce7d-49fb-a8e7-ed43970792bd-logs\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.940355 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16c9922f-a4a5-445f-a2b7-f252d25d1815-etc-machine-id\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.940980 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c9922f-a4a5-445f-a2b7-f252d25d1815-logs\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.941304 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdd1b143-6da9-463d-92a7-ff4c054e89fd-logs\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.946649 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data-custom\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.946701 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-scripts\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.947321 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-scripts\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.949122 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-scripts\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.949770 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data-custom\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.950138 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.950851 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.956391 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.959463 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data-custom\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.962355 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5thx\" (UniqueName: \"kubernetes.io/projected/0532be72-ce7d-49fb-a8e7-ed43970792bd-kube-api-access-z5thx\") pod \"cinder-api-2\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.965122 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq6f4\" (UniqueName: \"kubernetes.io/projected/16c9922f-a4a5-445f-a2b7-f252d25d1815-kube-api-access-bq6f4\") pod \"cinder-api-0\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:35 crc kubenswrapper[4989]: I1213 17:35:35.965253 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v7h8\" (UniqueName: \"kubernetes.io/projected/cdd1b143-6da9-463d-92a7-ff4c054e89fd-kube-api-access-6v7h8\") pod \"cinder-api-1\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:36 crc kubenswrapper[4989]: I1213 17:35:36.006066 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:36 crc kubenswrapper[4989]: I1213 17:35:36.025309 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f1acff6-4aa7-4860-aa71-298b67012ea4" path="/var/lib/kubelet/pods/4f1acff6-4aa7-4860-aa71-298b67012ea4/volumes" Dec 13 17:35:36 crc kubenswrapper[4989]: I1213 17:35:36.040161 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:36 crc kubenswrapper[4989]: I1213 17:35:36.050763 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:36 crc kubenswrapper[4989]: I1213 17:35:36.263105 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:35:36 crc kubenswrapper[4989]: I1213 17:35:36.533932 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 13 17:35:36 crc kubenswrapper[4989]: I1213 17:35:36.551207 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 13 17:35:36 crc kubenswrapper[4989]: W1213 17:35:36.558353 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdd1b143_6da9_463d_92a7_ff4c054e89fd.slice/crio-b833b6a1b766c649ca13d324678af7bb1d95cf5de890113e35d8820dae2de6d0 WatchSource:0}: Error finding container b833b6a1b766c649ca13d324678af7bb1d95cf5de890113e35d8820dae2de6d0: Status 404 returned error can't find the container with id b833b6a1b766c649ca13d324678af7bb1d95cf5de890113e35d8820dae2de6d0 Dec 13 17:35:36 crc kubenswrapper[4989]: W1213 17:35:36.558961 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0532be72_ce7d_49fb_a8e7_ed43970792bd.slice/crio-899f856b91433abf325f8c45c8979359e9add5f0b0480423440c035123aa7c01 WatchSource:0}: Error finding container 899f856b91433abf325f8c45c8979359e9add5f0b0480423440c035123aa7c01: Status 404 returned error can't find the container with id 899f856b91433abf325f8c45c8979359e9add5f0b0480423440c035123aa7c01 Dec 13 17:35:36 crc kubenswrapper[4989]: I1213 17:35:36.989142 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0532be72-ce7d-49fb-a8e7-ed43970792bd","Type":"ContainerStarted","Data":"899f856b91433abf325f8c45c8979359e9add5f0b0480423440c035123aa7c01"} Dec 13 17:35:36 crc kubenswrapper[4989]: I1213 17:35:36.990432 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"cdd1b143-6da9-463d-92a7-ff4c054e89fd","Type":"ContainerStarted","Data":"b833b6a1b766c649ca13d324678af7bb1d95cf5de890113e35d8820dae2de6d0"} Dec 13 17:35:36 crc kubenswrapper[4989]: I1213 17:35:36.992854 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"16c9922f-a4a5-445f-a2b7-f252d25d1815","Type":"ContainerStarted","Data":"444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861"} Dec 13 17:35:36 crc kubenswrapper[4989]: I1213 17:35:36.992913 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"16c9922f-a4a5-445f-a2b7-f252d25d1815","Type":"ContainerStarted","Data":"750cd28233bf6477c28f0a4b23ca7a937ce95e980f5c6a69a144905f30bdcdfc"} Dec 13 17:35:38 crc kubenswrapper[4989]: I1213 17:35:38.001966 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"cdd1b143-6da9-463d-92a7-ff4c054e89fd","Type":"ContainerStarted","Data":"ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d"} Dec 13 17:35:38 crc kubenswrapper[4989]: I1213 17:35:38.002367 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"cdd1b143-6da9-463d-92a7-ff4c054e89fd","Type":"ContainerStarted","Data":"7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc"} Dec 13 17:35:38 crc kubenswrapper[4989]: I1213 17:35:38.002383 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:38 crc kubenswrapper[4989]: I1213 17:35:38.004829 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"16c9922f-a4a5-445f-a2b7-f252d25d1815","Type":"ContainerStarted","Data":"df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726"} Dec 13 17:35:38 crc kubenswrapper[4989]: I1213 17:35:38.005103 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:38 crc kubenswrapper[4989]: I1213 17:35:38.007715 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0532be72-ce7d-49fb-a8e7-ed43970792bd","Type":"ContainerStarted","Data":"dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6"} Dec 13 17:35:38 crc kubenswrapper[4989]: I1213 17:35:38.007753 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0532be72-ce7d-49fb-a8e7-ed43970792bd","Type":"ContainerStarted","Data":"7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c"} Dec 13 17:35:38 crc kubenswrapper[4989]: I1213 17:35:38.008244 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:38 crc kubenswrapper[4989]: I1213 17:35:38.028358 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-1" podStartSLOduration=3.028338368 podStartE2EDuration="3.028338368s" podCreationTimestamp="2025-12-13 17:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:35:38.022870949 +0000 UTC m=+1152.629318107" watchObservedRunningTime="2025-12-13 17:35:38.028338368 +0000 UTC m=+1152.634785506" Dec 13 17:35:38 crc kubenswrapper[4989]: I1213 17:35:38.054679 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-2" podStartSLOduration=3.054658723 podStartE2EDuration="3.054658723s" podCreationTimestamp="2025-12-13 17:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:35:38.046361707 +0000 UTC m=+1152.652808845" watchObservedRunningTime="2025-12-13 17:35:38.054658723 +0000 UTC m=+1152.661105871" Dec 13 17:35:38 crc kubenswrapper[4989]: I1213 17:35:38.094627 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=3.094595569 podStartE2EDuration="3.094595569s" podCreationTimestamp="2025-12-13 17:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:35:38.083393692 +0000 UTC m=+1152.689840840" watchObservedRunningTime="2025-12-13 17:35:38.094595569 +0000 UTC m=+1152.701042747" Dec 13 17:35:44 crc kubenswrapper[4989]: I1213 17:35:44.015306 4989 scope.go:117] "RemoveContainer" containerID="3452a91b024721df6d1e71e9758645f089c80d3938d3dc800591f31be452eeda" Dec 13 17:35:44 crc kubenswrapper[4989]: I1213 17:35:44.015600 4989 scope.go:117] "RemoveContainer" containerID="a81daec09666018505c8871efbeb99f8f40e71e3acd5ecb94c2466d22bbb8dd8" Dec 13 17:35:44 crc kubenswrapper[4989]: E1213 17:35:44.015792 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:35:48 crc kubenswrapper[4989]: I1213 17:35:48.050343 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:48 crc kubenswrapper[4989]: I1213 17:35:48.121966 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:48 crc kubenswrapper[4989]: I1213 17:35:48.141233 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:49 crc kubenswrapper[4989]: I1213 17:35:49.092778 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 13 17:35:49 crc kubenswrapper[4989]: I1213 17:35:49.114150 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 13 17:35:49 crc kubenswrapper[4989]: I1213 17:35:49.121764 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-1" podUID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerName="cinder-api-log" containerID="cri-o://7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc" gracePeriod=30 Dec 13 17:35:49 crc kubenswrapper[4989]: I1213 17:35:49.121971 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-1" podUID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerName="cinder-api" containerID="cri-o://ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d" gracePeriod=30 Dec 13 17:35:49 crc kubenswrapper[4989]: I1213 17:35:49.122036 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-2" podUID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerName="cinder-api-log" containerID="cri-o://7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c" gracePeriod=30 Dec 13 17:35:49 crc kubenswrapper[4989]: I1213 17:35:49.122154 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-2" podUID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerName="cinder-api" containerID="cri-o://dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6" gracePeriod=30 Dec 13 17:35:49 crc kubenswrapper[4989]: I1213 17:35:49.132909 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-2" podUID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.99:8776/healthcheck\": EOF" Dec 13 17:35:49 crc kubenswrapper[4989]: I1213 17:35:49.135837 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-1" podUID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.100:8776/healthcheck\": EOF" Dec 13 17:35:50 crc kubenswrapper[4989]: I1213 17:35:50.129234 4989 generic.go:334] "Generic (PLEG): container finished" podID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerID="7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c" exitCode=143 Dec 13 17:35:50 crc kubenswrapper[4989]: I1213 17:35:50.129316 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0532be72-ce7d-49fb-a8e7-ed43970792bd","Type":"ContainerDied","Data":"7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c"} Dec 13 17:35:50 crc kubenswrapper[4989]: I1213 17:35:50.131569 4989 generic.go:334] "Generic (PLEG): container finished" podID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerID="7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc" exitCode=143 Dec 13 17:35:50 crc kubenswrapper[4989]: I1213 17:35:50.131595 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"cdd1b143-6da9-463d-92a7-ff4c054e89fd","Type":"ContainerDied","Data":"7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc"} Dec 13 17:35:53 crc kubenswrapper[4989]: I1213 17:35:53.539957 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-1" podUID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.100:8776/healthcheck\": read tcp 10.217.0.2:50066->10.217.0.100:8776: read: connection reset by peer" Dec 13 17:35:53 crc kubenswrapper[4989]: I1213 17:35:53.541712 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-2" podUID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.99:8776/healthcheck\": read tcp 10.217.0.2:34696->10.217.0.99:8776: read: connection reset by peer" Dec 13 17:35:53 crc kubenswrapper[4989]: I1213 17:35:53.948180 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.021239 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.059181 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-scripts\") pod \"0532be72-ce7d-49fb-a8e7-ed43970792bd\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.059247 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data\") pod \"0532be72-ce7d-49fb-a8e7-ed43970792bd\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.059296 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data-custom\") pod \"0532be72-ce7d-49fb-a8e7-ed43970792bd\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.059315 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5thx\" (UniqueName: \"kubernetes.io/projected/0532be72-ce7d-49fb-a8e7-ed43970792bd-kube-api-access-z5thx\") pod \"0532be72-ce7d-49fb-a8e7-ed43970792bd\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.059380 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0532be72-ce7d-49fb-a8e7-ed43970792bd-etc-machine-id\") pod \"0532be72-ce7d-49fb-a8e7-ed43970792bd\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.059395 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0532be72-ce7d-49fb-a8e7-ed43970792bd-logs\") pod \"0532be72-ce7d-49fb-a8e7-ed43970792bd\" (UID: \"0532be72-ce7d-49fb-a8e7-ed43970792bd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.059521 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0532be72-ce7d-49fb-a8e7-ed43970792bd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0532be72-ce7d-49fb-a8e7-ed43970792bd" (UID: "0532be72-ce7d-49fb-a8e7-ed43970792bd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.059726 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0532be72-ce7d-49fb-a8e7-ed43970792bd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.060054 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0532be72-ce7d-49fb-a8e7-ed43970792bd-logs" (OuterVolumeSpecName: "logs") pod "0532be72-ce7d-49fb-a8e7-ed43970792bd" (UID: "0532be72-ce7d-49fb-a8e7-ed43970792bd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.064577 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0532be72-ce7d-49fb-a8e7-ed43970792bd-kube-api-access-z5thx" (OuterVolumeSpecName: "kube-api-access-z5thx") pod "0532be72-ce7d-49fb-a8e7-ed43970792bd" (UID: "0532be72-ce7d-49fb-a8e7-ed43970792bd"). InnerVolumeSpecName "kube-api-access-z5thx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.064608 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0532be72-ce7d-49fb-a8e7-ed43970792bd" (UID: "0532be72-ce7d-49fb-a8e7-ed43970792bd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.065764 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-scripts" (OuterVolumeSpecName: "scripts") pod "0532be72-ce7d-49fb-a8e7-ed43970792bd" (UID: "0532be72-ce7d-49fb-a8e7-ed43970792bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.099179 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data" (OuterVolumeSpecName: "config-data") pod "0532be72-ce7d-49fb-a8e7-ed43970792bd" (UID: "0532be72-ce7d-49fb-a8e7-ed43970792bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.160555 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data-custom\") pod \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.160609 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdd1b143-6da9-463d-92a7-ff4c054e89fd-logs\") pod \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.160634 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6v7h8\" (UniqueName: \"kubernetes.io/projected/cdd1b143-6da9-463d-92a7-ff4c054e89fd-kube-api-access-6v7h8\") pod \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.160683 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-scripts\") pod \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.160811 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data\") pod \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.160879 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cdd1b143-6da9-463d-92a7-ff4c054e89fd-etc-machine-id\") pod \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\" (UID: \"cdd1b143-6da9-463d-92a7-ff4c054e89fd\") " Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.161178 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.161190 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.161199 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0532be72-ce7d-49fb-a8e7-ed43970792bd-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.161208 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5thx\" (UniqueName: \"kubernetes.io/projected/0532be72-ce7d-49fb-a8e7-ed43970792bd-kube-api-access-z5thx\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.161217 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0532be72-ce7d-49fb-a8e7-ed43970792bd-logs\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.161816 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cdd1b143-6da9-463d-92a7-ff4c054e89fd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cdd1b143-6da9-463d-92a7-ff4c054e89fd" (UID: "cdd1b143-6da9-463d-92a7-ff4c054e89fd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.162371 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdd1b143-6da9-463d-92a7-ff4c054e89fd-logs" (OuterVolumeSpecName: "logs") pod "cdd1b143-6da9-463d-92a7-ff4c054e89fd" (UID: "cdd1b143-6da9-463d-92a7-ff4c054e89fd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.163986 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cdd1b143-6da9-463d-92a7-ff4c054e89fd" (UID: "cdd1b143-6da9-463d-92a7-ff4c054e89fd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.164229 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-scripts" (OuterVolumeSpecName: "scripts") pod "cdd1b143-6da9-463d-92a7-ff4c054e89fd" (UID: "cdd1b143-6da9-463d-92a7-ff4c054e89fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.165279 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdd1b143-6da9-463d-92a7-ff4c054e89fd-kube-api-access-6v7h8" (OuterVolumeSpecName: "kube-api-access-6v7h8") pod "cdd1b143-6da9-463d-92a7-ff4c054e89fd" (UID: "cdd1b143-6da9-463d-92a7-ff4c054e89fd"). InnerVolumeSpecName "kube-api-access-6v7h8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.170563 4989 generic.go:334] "Generic (PLEG): container finished" podID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerID="dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6" exitCode=0 Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.170722 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0532be72-ce7d-49fb-a8e7-ed43970792bd","Type":"ContainerDied","Data":"dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6"} Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.170908 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0532be72-ce7d-49fb-a8e7-ed43970792bd","Type":"ContainerDied","Data":"899f856b91433abf325f8c45c8979359e9add5f0b0480423440c035123aa7c01"} Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.171046 4989 scope.go:117] "RemoveContainer" containerID="dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.171267 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.174630 4989 generic.go:334] "Generic (PLEG): container finished" podID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerID="ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d" exitCode=0 Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.174692 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"cdd1b143-6da9-463d-92a7-ff4c054e89fd","Type":"ContainerDied","Data":"ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d"} Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.174730 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.174734 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"cdd1b143-6da9-463d-92a7-ff4c054e89fd","Type":"ContainerDied","Data":"b833b6a1b766c649ca13d324678af7bb1d95cf5de890113e35d8820dae2de6d0"} Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.190079 4989 scope.go:117] "RemoveContainer" containerID="7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.211177 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.217836 4989 scope.go:117] "RemoveContainer" containerID="dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.218422 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data" (OuterVolumeSpecName: "config-data") pod "cdd1b143-6da9-463d-92a7-ff4c054e89fd" (UID: "cdd1b143-6da9-463d-92a7-ff4c054e89fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:54 crc kubenswrapper[4989]: E1213 17:35:54.218670 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6\": container with ID starting with dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6 not found: ID does not exist" containerID="dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.218730 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6"} err="failed to get container status \"dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6\": rpc error: code = NotFound desc = could not find container \"dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6\": container with ID starting with dcb9817d789b9d54662df4a88dc449c1ecf00513651a445049757176981322c6 not found: ID does not exist" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.218757 4989 scope.go:117] "RemoveContainer" containerID="7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c" Dec 13 17:35:54 crc kubenswrapper[4989]: E1213 17:35:54.219303 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c\": container with ID starting with 7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c not found: ID does not exist" containerID="7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.219354 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c"} err="failed to get container status \"7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c\": rpc error: code = NotFound desc = could not find container \"7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c\": container with ID starting with 7d3104a32e17ec67170d823d192efa06d4725e78093740524f913d5ba9c1bc2c not found: ID does not exist" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.219389 4989 scope.go:117] "RemoveContainer" containerID="ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.222930 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.238401 4989 scope.go:117] "RemoveContainer" containerID="7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.251783 4989 scope.go:117] "RemoveContainer" containerID="ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d" Dec 13 17:35:54 crc kubenswrapper[4989]: E1213 17:35:54.253058 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d\": container with ID starting with ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d not found: ID does not exist" containerID="ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.253090 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d"} err="failed to get container status \"ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d\": rpc error: code = NotFound desc = could not find container \"ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d\": container with ID starting with ed09b3c44897a536b7b1d04f7c094f26f909da08478d9c8c7ea68944899a647d not found: ID does not exist" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.253110 4989 scope.go:117] "RemoveContainer" containerID="7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc" Dec 13 17:35:54 crc kubenswrapper[4989]: E1213 17:35:54.253589 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc\": container with ID starting with 7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc not found: ID does not exist" containerID="7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.253665 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc"} err="failed to get container status \"7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc\": rpc error: code = NotFound desc = could not find container \"7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc\": container with ID starting with 7b4287f85b5c3a5e9f6984d71cadc214528c4a1a69e0232d56e75d0ed6103fdc not found: ID does not exist" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.262443 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cdd1b143-6da9-463d-92a7-ff4c054e89fd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.262475 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.262488 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdd1b143-6da9-463d-92a7-ff4c054e89fd-logs\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.262502 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6v7h8\" (UniqueName: \"kubernetes.io/projected/cdd1b143-6da9-463d-92a7-ff4c054e89fd-kube-api-access-6v7h8\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.262514 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.262524 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdd1b143-6da9-463d-92a7-ff4c054e89fd-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.509293 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 13 17:35:54 crc kubenswrapper[4989]: I1213 17:35:54.516330 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.015098 4989 scope.go:117] "RemoveContainer" containerID="3452a91b024721df6d1e71e9758645f089c80d3938d3dc800591f31be452eeda" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.015424 4989 scope.go:117] "RemoveContainer" containerID="a81daec09666018505c8871efbeb99f8f40e71e3acd5ecb94c2466d22bbb8dd8" Dec 13 17:35:55 crc kubenswrapper[4989]: E1213 17:35:55.015898 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(aea7e175-a80a-4630-9b10-278591652873)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="aea7e175-a80a-4630-9b10-278591652873" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.409525 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-2k9tr"] Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.420269 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-2k9tr"] Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.446507 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.446813 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="0a999f01-db9b-4b04-9b64-1430f1e7eac1" containerName="cinder-backup" containerID="cri-o://f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56" gracePeriod=30 Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.446856 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="0a999f01-db9b-4b04-9b64-1430f1e7eac1" containerName="probe" containerID="cri-o://a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5" gracePeriod=30 Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.455077 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.455598 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="b4cccfb2-6a7a-456b-a355-1300d7f59fef" containerName="cinder-scheduler" containerID="cri-o://86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224" gracePeriod=30 Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.456225 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="b4cccfb2-6a7a-456b-a355-1300d7f59fef" containerName="probe" containerID="cri-o://e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff" gracePeriod=30 Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.494963 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder2ef4-account-delete-hf46z"] Dec 13 17:35:55 crc kubenswrapper[4989]: E1213 17:35:55.495773 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerName="cinder-api" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.495873 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerName="cinder-api" Dec 13 17:35:55 crc kubenswrapper[4989]: E1213 17:35:55.495969 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerName="cinder-api" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.496030 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerName="cinder-api" Dec 13 17:35:55 crc kubenswrapper[4989]: E1213 17:35:55.496097 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerName="cinder-api-log" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.496149 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerName="cinder-api-log" Dec 13 17:35:55 crc kubenswrapper[4989]: E1213 17:35:55.496213 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerName="cinder-api-log" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.496264 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerName="cinder-api-log" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.496677 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerName="cinder-api-log" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.496775 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerName="cinder-api" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.496883 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0532be72-ce7d-49fb-a8e7-ed43970792bd" containerName="cinder-api-log" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.496940 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" containerName="cinder-api" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.497717 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.522392 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder2ef4-account-delete-hf46z"] Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.529282 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.573767 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.574057 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="16c9922f-a4a5-445f-a2b7-f252d25d1815" containerName="cinder-api-log" containerID="cri-o://444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861" gracePeriod=30 Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.574200 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="16c9922f-a4a5-445f-a2b7-f252d25d1815" containerName="cinder-api" containerID="cri-o://df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726" gracePeriod=30 Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.585592 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtq8g\" (UniqueName: \"kubernetes.io/projected/d3419168-fb01-4821-b723-79aa7712cca4-kube-api-access-mtq8g\") pod \"cinder2ef4-account-delete-hf46z\" (UID: \"d3419168-fb01-4821-b723-79aa7712cca4\") " pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.585723 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3419168-fb01-4821-b723-79aa7712cca4-operator-scripts\") pod \"cinder2ef4-account-delete-hf46z\" (UID: \"d3419168-fb01-4821-b723-79aa7712cca4\") " pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.686935 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtq8g\" (UniqueName: \"kubernetes.io/projected/d3419168-fb01-4821-b723-79aa7712cca4-kube-api-access-mtq8g\") pod \"cinder2ef4-account-delete-hf46z\" (UID: \"d3419168-fb01-4821-b723-79aa7712cca4\") " pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.687083 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3419168-fb01-4821-b723-79aa7712cca4-operator-scripts\") pod \"cinder2ef4-account-delete-hf46z\" (UID: \"d3419168-fb01-4821-b723-79aa7712cca4\") " pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.688352 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3419168-fb01-4821-b723-79aa7712cca4-operator-scripts\") pod \"cinder2ef4-account-delete-hf46z\" (UID: \"d3419168-fb01-4821-b723-79aa7712cca4\") " pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.729397 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtq8g\" (UniqueName: \"kubernetes.io/projected/d3419168-fb01-4821-b723-79aa7712cca4-kube-api-access-mtq8g\") pod \"cinder2ef4-account-delete-hf46z\" (UID: \"d3419168-fb01-4821-b723-79aa7712cca4\") " pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.816256 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" Dec 13 17:35:55 crc kubenswrapper[4989]: I1213 17:35:55.942822 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.025603 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0532be72-ce7d-49fb-a8e7-ed43970792bd" path="/var/lib/kubelet/pods/0532be72-ce7d-49fb-a8e7-ed43970792bd/volumes" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.026299 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54e8ac0e-d364-4244-a054-8f213285bacb" path="/var/lib/kubelet/pods/54e8ac0e-d364-4244-a054-8f213285bacb/volumes" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.026843 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdd1b143-6da9-463d-92a7-ff4c054e89fd" path="/var/lib/kubelet/pods/cdd1b143-6da9-463d-92a7-ff4c054e89fd/volumes" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093246 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-lib-modules\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093286 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-nvme\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093320 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data-custom\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093360 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-iscsi\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093377 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-dev\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093393 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093432 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-brick\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093455 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-scripts\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093472 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-dev" (OuterVolumeSpecName: "dev") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093479 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdv5t\" (UniqueName: \"kubernetes.io/projected/aea7e175-a80a-4630-9b10-278591652873-kube-api-access-bdv5t\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093541 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093587 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093541 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093561 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-cinder\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093649 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-run\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093687 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-machine-id\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093771 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-run" (OuterVolumeSpecName: "run") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093780 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093832 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093902 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-sys\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093931 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-lib-cinder\") pod \"aea7e175-a80a-4630-9b10-278591652873\" (UID: \"aea7e175-a80a-4630-9b10-278591652873\") " Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.093993 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-sys" (OuterVolumeSpecName: "sys") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.094116 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.094245 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.094918 4989 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-sys\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.094936 4989 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.094950 4989 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.094961 4989 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.094970 4989 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.094979 4989 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-dev\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.094988 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.094997 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.095006 4989 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-run\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.095018 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aea7e175-a80a-4630-9b10-278591652873-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.110063 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.132627 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aea7e175-a80a-4630-9b10-278591652873-kube-api-access-bdv5t" (OuterVolumeSpecName: "kube-api-access-bdv5t") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "kube-api-access-bdv5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.135919 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-scripts" (OuterVolumeSpecName: "scripts") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.157814 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data" (OuterVolumeSpecName: "config-data") pod "aea7e175-a80a-4630-9b10-278591652873" (UID: "aea7e175-a80a-4630-9b10-278591652873"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.195850 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.195890 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.195905 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdv5t\" (UniqueName: \"kubernetes.io/projected/aea7e175-a80a-4630-9b10-278591652873-kube-api-access-bdv5t\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.195917 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea7e175-a80a-4630-9b10-278591652873-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.196965 4989 generic.go:334] "Generic (PLEG): container finished" podID="0a999f01-db9b-4b04-9b64-1430f1e7eac1" containerID="a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5" exitCode=0 Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.197009 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0a999f01-db9b-4b04-9b64-1430f1e7eac1","Type":"ContainerDied","Data":"a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5"} Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.198676 4989 generic.go:334] "Generic (PLEG): container finished" podID="16c9922f-a4a5-445f-a2b7-f252d25d1815" containerID="444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861" exitCode=143 Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.198746 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"16c9922f-a4a5-445f-a2b7-f252d25d1815","Type":"ContainerDied","Data":"444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861"} Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.200694 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"aea7e175-a80a-4630-9b10-278591652873","Type":"ContainerDied","Data":"5be0815972e3a4cd63f94a9d0df1fb8fb477e7d10dcb528353fb492239737e36"} Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.200723 4989 scope.go:117] "RemoveContainer" containerID="a81daec09666018505c8871efbeb99f8f40e71e3acd5ecb94c2466d22bbb8dd8" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.200769 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.242071 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.248203 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.250065 4989 scope.go:117] "RemoveContainer" containerID="3452a91b024721df6d1e71e9758645f089c80d3938d3dc800591f31be452eeda" Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.284624 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder2ef4-account-delete-hf46z"] Dec 13 17:35:56 crc kubenswrapper[4989]: W1213 17:35:56.287094 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3419168_fb01_4821_b723_79aa7712cca4.slice/crio-f3c9e44386b88e1b7ed153adc0684017aadc90d11bd20adb5e68c8b80a2ad822 WatchSource:0}: Error finding container f3c9e44386b88e1b7ed153adc0684017aadc90d11bd20adb5e68c8b80a2ad822: Status 404 returned error can't find the container with id f3c9e44386b88e1b7ed153adc0684017aadc90d11bd20adb5e68c8b80a2ad822 Dec 13 17:35:56 crc kubenswrapper[4989]: I1213 17:35:56.875408 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.012158 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data\") pod \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.012254 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data-custom\") pod \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.012301 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b4cccfb2-6a7a-456b-a355-1300d7f59fef-etc-machine-id\") pod \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.012355 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-scripts\") pod \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.012434 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgrrb\" (UniqueName: \"kubernetes.io/projected/b4cccfb2-6a7a-456b-a355-1300d7f59fef-kube-api-access-fgrrb\") pod \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\" (UID: \"b4cccfb2-6a7a-456b-a355-1300d7f59fef\") " Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.012420 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4cccfb2-6a7a-456b-a355-1300d7f59fef-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b4cccfb2-6a7a-456b-a355-1300d7f59fef" (UID: "b4cccfb2-6a7a-456b-a355-1300d7f59fef"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.012726 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b4cccfb2-6a7a-456b-a355-1300d7f59fef-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.017486 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4cccfb2-6a7a-456b-a355-1300d7f59fef-kube-api-access-fgrrb" (OuterVolumeSpecName: "kube-api-access-fgrrb") pod "b4cccfb2-6a7a-456b-a355-1300d7f59fef" (UID: "b4cccfb2-6a7a-456b-a355-1300d7f59fef"). InnerVolumeSpecName "kube-api-access-fgrrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.017663 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-scripts" (OuterVolumeSpecName: "scripts") pod "b4cccfb2-6a7a-456b-a355-1300d7f59fef" (UID: "b4cccfb2-6a7a-456b-a355-1300d7f59fef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.032006 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b4cccfb2-6a7a-456b-a355-1300d7f59fef" (UID: "b4cccfb2-6a7a-456b-a355-1300d7f59fef"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.089933 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data" (OuterVolumeSpecName: "config-data") pod "b4cccfb2-6a7a-456b-a355-1300d7f59fef" (UID: "b4cccfb2-6a7a-456b-a355-1300d7f59fef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.114466 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.114570 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgrrb\" (UniqueName: \"kubernetes.io/projected/b4cccfb2-6a7a-456b-a355-1300d7f59fef-kube-api-access-fgrrb\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.114589 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.114599 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b4cccfb2-6a7a-456b-a355-1300d7f59fef-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.212636 4989 generic.go:334] "Generic (PLEG): container finished" podID="b4cccfb2-6a7a-456b-a355-1300d7f59fef" containerID="e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff" exitCode=0 Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.212667 4989 generic.go:334] "Generic (PLEG): container finished" podID="b4cccfb2-6a7a-456b-a355-1300d7f59fef" containerID="86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224" exitCode=0 Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.212725 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.212737 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"b4cccfb2-6a7a-456b-a355-1300d7f59fef","Type":"ContainerDied","Data":"e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff"} Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.212856 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"b4cccfb2-6a7a-456b-a355-1300d7f59fef","Type":"ContainerDied","Data":"86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224"} Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.212927 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"b4cccfb2-6a7a-456b-a355-1300d7f59fef","Type":"ContainerDied","Data":"88459dad79e652faee46a01d91b9d33de65c2064940b0a40fae40f9a3db621a1"} Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.212903 4989 scope.go:117] "RemoveContainer" containerID="e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.214397 4989 generic.go:334] "Generic (PLEG): container finished" podID="d3419168-fb01-4821-b723-79aa7712cca4" containerID="8180eb9a5e8b9879718f47a629137ef23241a2b78ec10aa98a03edaa37cbe592" exitCode=0 Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.214430 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" event={"ID":"d3419168-fb01-4821-b723-79aa7712cca4","Type":"ContainerDied","Data":"8180eb9a5e8b9879718f47a629137ef23241a2b78ec10aa98a03edaa37cbe592"} Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.214482 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" event={"ID":"d3419168-fb01-4821-b723-79aa7712cca4","Type":"ContainerStarted","Data":"f3c9e44386b88e1b7ed153adc0684017aadc90d11bd20adb5e68c8b80a2ad822"} Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.251649 4989 scope.go:117] "RemoveContainer" containerID="86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.274196 4989 scope.go:117] "RemoveContainer" containerID="e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff" Dec 13 17:35:57 crc kubenswrapper[4989]: E1213 17:35:57.313125 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff\": container with ID starting with e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff not found: ID does not exist" containerID="e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.313185 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff"} err="failed to get container status \"e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff\": rpc error: code = NotFound desc = could not find container \"e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff\": container with ID starting with e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff not found: ID does not exist" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.313218 4989 scope.go:117] "RemoveContainer" containerID="86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224" Dec 13 17:35:57 crc kubenswrapper[4989]: E1213 17:35:57.317404 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224\": container with ID starting with 86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224 not found: ID does not exist" containerID="86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.317574 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224"} err="failed to get container status \"86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224\": rpc error: code = NotFound desc = could not find container \"86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224\": container with ID starting with 86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224 not found: ID does not exist" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.317728 4989 scope.go:117] "RemoveContainer" containerID="e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.318564 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff"} err="failed to get container status \"e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff\": rpc error: code = NotFound desc = could not find container \"e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff\": container with ID starting with e74216b05c537d5610f44244c0da386486826ea8fd1a4773fe79adb297814fff not found: ID does not exist" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.318616 4989 scope.go:117] "RemoveContainer" containerID="86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.319091 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224"} err="failed to get container status \"86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224\": rpc error: code = NotFound desc = could not find container \"86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224\": container with ID starting with 86bf75b64ae0e906445f4fbeaa4dce62d7de960e549b3519dd3ef35a93084224 not found: ID does not exist" Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.325282 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:35:57 crc kubenswrapper[4989]: I1213 17:35:57.332758 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:35:58 crc kubenswrapper[4989]: I1213 17:35:58.022417 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aea7e175-a80a-4630-9b10-278591652873" path="/var/lib/kubelet/pods/aea7e175-a80a-4630-9b10-278591652873/volumes" Dec 13 17:35:58 crc kubenswrapper[4989]: I1213 17:35:58.023211 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4cccfb2-6a7a-456b-a355-1300d7f59fef" path="/var/lib/kubelet/pods/b4cccfb2-6a7a-456b-a355-1300d7f59fef/volumes" Dec 13 17:35:58 crc kubenswrapper[4989]: I1213 17:35:58.555940 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" Dec 13 17:35:58 crc kubenswrapper[4989]: I1213 17:35:58.729716 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="16c9922f-a4a5-445f-a2b7-f252d25d1815" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.98:8776/healthcheck\": read tcp 10.217.0.2:33836->10.217.0.98:8776: read: connection reset by peer" Dec 13 17:35:58 crc kubenswrapper[4989]: I1213 17:35:58.741173 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtq8g\" (UniqueName: \"kubernetes.io/projected/d3419168-fb01-4821-b723-79aa7712cca4-kube-api-access-mtq8g\") pod \"d3419168-fb01-4821-b723-79aa7712cca4\" (UID: \"d3419168-fb01-4821-b723-79aa7712cca4\") " Dec 13 17:35:58 crc kubenswrapper[4989]: I1213 17:35:58.741384 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3419168-fb01-4821-b723-79aa7712cca4-operator-scripts\") pod \"d3419168-fb01-4821-b723-79aa7712cca4\" (UID: \"d3419168-fb01-4821-b723-79aa7712cca4\") " Dec 13 17:35:58 crc kubenswrapper[4989]: I1213 17:35:58.742074 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3419168-fb01-4821-b723-79aa7712cca4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3419168-fb01-4821-b723-79aa7712cca4" (UID: "d3419168-fb01-4821-b723-79aa7712cca4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:35:58 crc kubenswrapper[4989]: I1213 17:35:58.747710 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3419168-fb01-4821-b723-79aa7712cca4-kube-api-access-mtq8g" (OuterVolumeSpecName: "kube-api-access-mtq8g") pod "d3419168-fb01-4821-b723-79aa7712cca4" (UID: "d3419168-fb01-4821-b723-79aa7712cca4"). InnerVolumeSpecName "kube-api-access-mtq8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:35:58 crc kubenswrapper[4989]: I1213 17:35:58.846216 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3419168-fb01-4821-b723-79aa7712cca4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:58 crc kubenswrapper[4989]: I1213 17:35:58.846280 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtq8g\" (UniqueName: \"kubernetes.io/projected/d3419168-fb01-4821-b723-79aa7712cca4-kube-api-access-mtq8g\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.089347 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.232243 4989 generic.go:334] "Generic (PLEG): container finished" podID="16c9922f-a4a5-445f-a2b7-f252d25d1815" containerID="df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726" exitCode=0 Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.232329 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"16c9922f-a4a5-445f-a2b7-f252d25d1815","Type":"ContainerDied","Data":"df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726"} Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.232632 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"16c9922f-a4a5-445f-a2b7-f252d25d1815","Type":"ContainerDied","Data":"750cd28233bf6477c28f0a4b23ca7a937ce95e980f5c6a69a144905f30bdcdfc"} Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.232654 4989 scope.go:117] "RemoveContainer" containerID="df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.232392 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.237543 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" event={"ID":"d3419168-fb01-4821-b723-79aa7712cca4","Type":"ContainerDied","Data":"f3c9e44386b88e1b7ed153adc0684017aadc90d11bd20adb5e68c8b80a2ad822"} Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.237589 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3c9e44386b88e1b7ed153adc0684017aadc90d11bd20adb5e68c8b80a2ad822" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.237601 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder2ef4-account-delete-hf46z" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.250880 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bq6f4\" (UniqueName: \"kubernetes.io/projected/16c9922f-a4a5-445f-a2b7-f252d25d1815-kube-api-access-bq6f4\") pod \"16c9922f-a4a5-445f-a2b7-f252d25d1815\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.250989 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16c9922f-a4a5-445f-a2b7-f252d25d1815-etc-machine-id\") pod \"16c9922f-a4a5-445f-a2b7-f252d25d1815\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.251034 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-scripts\") pod \"16c9922f-a4a5-445f-a2b7-f252d25d1815\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.251133 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data\") pod \"16c9922f-a4a5-445f-a2b7-f252d25d1815\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.251294 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c9922f-a4a5-445f-a2b7-f252d25d1815-logs\") pod \"16c9922f-a4a5-445f-a2b7-f252d25d1815\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.251296 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/16c9922f-a4a5-445f-a2b7-f252d25d1815-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "16c9922f-a4a5-445f-a2b7-f252d25d1815" (UID: "16c9922f-a4a5-445f-a2b7-f252d25d1815"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.251351 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data-custom\") pod \"16c9922f-a4a5-445f-a2b7-f252d25d1815\" (UID: \"16c9922f-a4a5-445f-a2b7-f252d25d1815\") " Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.251756 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16c9922f-a4a5-445f-a2b7-f252d25d1815-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.252262 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16c9922f-a4a5-445f-a2b7-f252d25d1815-logs" (OuterVolumeSpecName: "logs") pod "16c9922f-a4a5-445f-a2b7-f252d25d1815" (UID: "16c9922f-a4a5-445f-a2b7-f252d25d1815"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.255371 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16c9922f-a4a5-445f-a2b7-f252d25d1815-kube-api-access-bq6f4" (OuterVolumeSpecName: "kube-api-access-bq6f4") pod "16c9922f-a4a5-445f-a2b7-f252d25d1815" (UID: "16c9922f-a4a5-445f-a2b7-f252d25d1815"). InnerVolumeSpecName "kube-api-access-bq6f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.258350 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "16c9922f-a4a5-445f-a2b7-f252d25d1815" (UID: "16c9922f-a4a5-445f-a2b7-f252d25d1815"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.267266 4989 scope.go:117] "RemoveContainer" containerID="444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.270955 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-scripts" (OuterVolumeSpecName: "scripts") pod "16c9922f-a4a5-445f-a2b7-f252d25d1815" (UID: "16c9922f-a4a5-445f-a2b7-f252d25d1815"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.302185 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data" (OuterVolumeSpecName: "config-data") pod "16c9922f-a4a5-445f-a2b7-f252d25d1815" (UID: "16c9922f-a4a5-445f-a2b7-f252d25d1815"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.340956 4989 scope.go:117] "RemoveContainer" containerID="df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726" Dec 13 17:35:59 crc kubenswrapper[4989]: E1213 17:35:59.341297 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726\": container with ID starting with df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726 not found: ID does not exist" containerID="df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.341325 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726"} err="failed to get container status \"df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726\": rpc error: code = NotFound desc = could not find container \"df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726\": container with ID starting with df6c8775760139bc0d948ecc4ef47faf588d382e1453a987906968f6d3009726 not found: ID does not exist" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.341343 4989 scope.go:117] "RemoveContainer" containerID="444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861" Dec 13 17:35:59 crc kubenswrapper[4989]: E1213 17:35:59.341510 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861\": container with ID starting with 444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861 not found: ID does not exist" containerID="444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.341530 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861"} err="failed to get container status \"444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861\": rpc error: code = NotFound desc = could not find container \"444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861\": container with ID starting with 444f79dc24359a0e15026d95ba6e209b0cdbf33202fe3864e0cc794d5cf48861 not found: ID does not exist" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.353258 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.353285 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.353297 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16c9922f-a4a5-445f-a2b7-f252d25d1815-logs\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.353308 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16c9922f-a4a5-445f-a2b7-f252d25d1815-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.353318 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bq6f4\" (UniqueName: \"kubernetes.io/projected/16c9922f-a4a5-445f-a2b7-f252d25d1815-kube-api-access-bq6f4\") on node \"crc\" DevicePath \"\"" Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.567380 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:35:59 crc kubenswrapper[4989]: I1213 17:35:59.574910 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.024138 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16c9922f-a4a5-445f-a2b7-f252d25d1815" path="/var/lib/kubelet/pods/16c9922f-a4a5-445f-a2b7-f252d25d1815/volumes" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.513212 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-ws7zm"] Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.523028 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-ws7zm"] Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.531319 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch"] Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.537996 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder2ef4-account-delete-hf46z"] Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.542934 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-2ef4-account-create-update-tp7ch"] Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.547687 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder2ef4-account-delete-hf46z"] Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.657146 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.727041 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-create-78cjs"] Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729586 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4cccfb2-6a7a-456b-a355-1300d7f59fef" containerName="cinder-scheduler" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729613 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4cccfb2-6a7a-456b-a355-1300d7f59fef" containerName="cinder-scheduler" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729622 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729629 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729644 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c9922f-a4a5-445f-a2b7-f252d25d1815" containerName="cinder-api" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729652 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c9922f-a4a5-445f-a2b7-f252d25d1815" containerName="cinder-api" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729665 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3419168-fb01-4821-b723-79aa7712cca4" containerName="mariadb-account-delete" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729673 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3419168-fb01-4821-b723-79aa7712cca4" containerName="mariadb-account-delete" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729684 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4cccfb2-6a7a-456b-a355-1300d7f59fef" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729691 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4cccfb2-6a7a-456b-a355-1300d7f59fef" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729700 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729708 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729717 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16c9922f-a4a5-445f-a2b7-f252d25d1815" containerName="cinder-api-log" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729725 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="16c9922f-a4a5-445f-a2b7-f252d25d1815" containerName="cinder-api-log" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729736 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a999f01-db9b-4b04-9b64-1430f1e7eac1" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729743 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a999f01-db9b-4b04-9b64-1430f1e7eac1" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729757 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729764 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729776 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a999f01-db9b-4b04-9b64-1430f1e7eac1" containerName="cinder-backup" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729783 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a999f01-db9b-4b04-9b64-1430f1e7eac1" containerName="cinder-backup" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729815 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729822 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729832 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729840 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.729851 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.729858 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730016 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730029 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a999f01-db9b-4b04-9b64-1430f1e7eac1" containerName="cinder-backup" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730041 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3419168-fb01-4821-b723-79aa7712cca4" containerName="mariadb-account-delete" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730049 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730061 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a999f01-db9b-4b04-9b64-1430f1e7eac1" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730072 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c9922f-a4a5-445f-a2b7-f252d25d1815" containerName="cinder-api" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730083 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730092 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4cccfb2-6a7a-456b-a355-1300d7f59fef" containerName="cinder-scheduler" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730101 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730110 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730120 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="16c9922f-a4a5-445f-a2b7-f252d25d1815" containerName="cinder-api-log" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730130 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730140 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4cccfb2-6a7a-456b-a355-1300d7f59fef" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.730753 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-78cjs" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.734027 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-8848-account-create-update-q84vz"] Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.734394 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.734412 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: E1213 17:36:00.734443 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.734452 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.734596 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="cinder-volume" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.734616 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea7e175-a80a-4630-9b10-278591652873" containerName="probe" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.735204 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.739174 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-78cjs"] Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.739274 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-db-secret" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.746091 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-8848-account-create-update-q84vz"] Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773484 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-brick\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773527 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-run\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773558 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-scripts\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773581 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wbcx\" (UniqueName: \"kubernetes.io/projected/0a999f01-db9b-4b04-9b64-1430f1e7eac1-kube-api-access-6wbcx\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773625 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data-custom\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773690 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-machine-id\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773716 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773740 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-sys\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773760 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-nvme\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773803 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-iscsi\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773823 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-lib-cinder\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773847 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-cinder\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773898 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-dev\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773912 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773925 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773938 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-lib-modules\") pod \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\" (UID: \"0a999f01-db9b-4b04-9b64-1430f1e7eac1\") " Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.773975 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774015 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774016 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774041 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-dev" (OuterVolumeSpecName: "dev") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774113 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774541 4989 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774563 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774578 4989 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-dev\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774591 4989 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774602 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774613 4989 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774624 4989 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774658 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-sys" (OuterVolumeSpecName: "sys") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774694 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.774720 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-run" (OuterVolumeSpecName: "run") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.779236 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-scripts" (OuterVolumeSpecName: "scripts") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.779588 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a999f01-db9b-4b04-9b64-1430f1e7eac1-kube-api-access-6wbcx" (OuterVolumeSpecName: "kube-api-access-6wbcx") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "kube-api-access-6wbcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.788628 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.840901 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data" (OuterVolumeSpecName: "config-data") pod "0a999f01-db9b-4b04-9b64-1430f1e7eac1" (UID: "0a999f01-db9b-4b04-9b64-1430f1e7eac1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.876400 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27766e3d-62dd-4c50-85d5-e4ae824acd54-operator-scripts\") pod \"cinder-db-create-78cjs\" (UID: \"27766e3d-62dd-4c50-85d5-e4ae824acd54\") " pod="cinder-kuttl-tests/cinder-db-create-78cjs" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.876468 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-operator-scripts\") pod \"cinder-8848-account-create-update-q84vz\" (UID: \"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d\") " pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.876499 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz86j\" (UniqueName: \"kubernetes.io/projected/27766e3d-62dd-4c50-85d5-e4ae824acd54-kube-api-access-tz86j\") pod \"cinder-db-create-78cjs\" (UID: \"27766e3d-62dd-4c50-85d5-e4ae824acd54\") " pod="cinder-kuttl-tests/cinder-db-create-78cjs" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.876702 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79wmv\" (UniqueName: \"kubernetes.io/projected/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-kube-api-access-79wmv\") pod \"cinder-8848-account-create-update-q84vz\" (UID: \"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d\") " pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.876950 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.876977 4989 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-run\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.877000 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wbcx\" (UniqueName: \"kubernetes.io/projected/0a999f01-db9b-4b04-9b64-1430f1e7eac1-kube-api-access-6wbcx\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.877021 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.877040 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.877061 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a999f01-db9b-4b04-9b64-1430f1e7eac1-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.877073 4989 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0a999f01-db9b-4b04-9b64-1430f1e7eac1-sys\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.978236 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79wmv\" (UniqueName: \"kubernetes.io/projected/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-kube-api-access-79wmv\") pod \"cinder-8848-account-create-update-q84vz\" (UID: \"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d\") " pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.978322 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27766e3d-62dd-4c50-85d5-e4ae824acd54-operator-scripts\") pod \"cinder-db-create-78cjs\" (UID: \"27766e3d-62dd-4c50-85d5-e4ae824acd54\") " pod="cinder-kuttl-tests/cinder-db-create-78cjs" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.978372 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-operator-scripts\") pod \"cinder-8848-account-create-update-q84vz\" (UID: \"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d\") " pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.978405 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz86j\" (UniqueName: \"kubernetes.io/projected/27766e3d-62dd-4c50-85d5-e4ae824acd54-kube-api-access-tz86j\") pod \"cinder-db-create-78cjs\" (UID: \"27766e3d-62dd-4c50-85d5-e4ae824acd54\") " pod="cinder-kuttl-tests/cinder-db-create-78cjs" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.979535 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27766e3d-62dd-4c50-85d5-e4ae824acd54-operator-scripts\") pod \"cinder-db-create-78cjs\" (UID: \"27766e3d-62dd-4c50-85d5-e4ae824acd54\") " pod="cinder-kuttl-tests/cinder-db-create-78cjs" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.981167 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-operator-scripts\") pod \"cinder-8848-account-create-update-q84vz\" (UID: \"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d\") " pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.995429 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz86j\" (UniqueName: \"kubernetes.io/projected/27766e3d-62dd-4c50-85d5-e4ae824acd54-kube-api-access-tz86j\") pod \"cinder-db-create-78cjs\" (UID: \"27766e3d-62dd-4c50-85d5-e4ae824acd54\") " pod="cinder-kuttl-tests/cinder-db-create-78cjs" Dec 13 17:36:00 crc kubenswrapper[4989]: I1213 17:36:00.996952 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79wmv\" (UniqueName: \"kubernetes.io/projected/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-kube-api-access-79wmv\") pod \"cinder-8848-account-create-update-q84vz\" (UID: \"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d\") " pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.111355 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-78cjs" Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.170742 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.256350 4989 generic.go:334] "Generic (PLEG): container finished" podID="0a999f01-db9b-4b04-9b64-1430f1e7eac1" containerID="f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56" exitCode=0 Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.256559 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0a999f01-db9b-4b04-9b64-1430f1e7eac1","Type":"ContainerDied","Data":"f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56"} Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.256973 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0a999f01-db9b-4b04-9b64-1430f1e7eac1","Type":"ContainerDied","Data":"b08d8efbe70d67f14d03dfaacfbd8bdb0c2b1f04b7d884aa184cc930424f0c7d"} Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.256653 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.257005 4989 scope.go:117] "RemoveContainer" containerID="a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5" Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.279941 4989 scope.go:117] "RemoveContainer" containerID="f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56" Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.301592 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.311381 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.319215 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-78cjs"] Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.323593 4989 scope.go:117] "RemoveContainer" containerID="a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5" Dec 13 17:36:01 crc kubenswrapper[4989]: E1213 17:36:01.323981 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5\": container with ID starting with a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5 not found: ID does not exist" containerID="a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5" Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.324012 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5"} err="failed to get container status \"a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5\": rpc error: code = NotFound desc = could not find container \"a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5\": container with ID starting with a7759341248a31eecd59b0b97703dc76795a292c12179308d519190cb3e3e3e5 not found: ID does not exist" Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.324037 4989 scope.go:117] "RemoveContainer" containerID="f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56" Dec 13 17:36:01 crc kubenswrapper[4989]: E1213 17:36:01.324287 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56\": container with ID starting with f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56 not found: ID does not exist" containerID="f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56" Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.324309 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56"} err="failed to get container status \"f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56\": rpc error: code = NotFound desc = could not find container \"f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56\": container with ID starting with f56d55b53eeb6e14e0d82f8c94dd67668af272e6082b8579267c0297d7884e56 not found: ID does not exist" Dec 13 17:36:01 crc kubenswrapper[4989]: W1213 17:36:01.327819 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27766e3d_62dd_4c50_85d5_e4ae824acd54.slice/crio-43e9025a43ded95658769fb9e1fe782f972eab02b20190c7d49bfbdf5d81edf0 WatchSource:0}: Error finding container 43e9025a43ded95658769fb9e1fe782f972eab02b20190c7d49bfbdf5d81edf0: Status 404 returned error can't find the container with id 43e9025a43ded95658769fb9e1fe782f972eab02b20190c7d49bfbdf5d81edf0 Dec 13 17:36:01 crc kubenswrapper[4989]: I1213 17:36:01.597160 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-8848-account-create-update-q84vz"] Dec 13 17:36:01 crc kubenswrapper[4989]: W1213 17:36:01.606724 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfb45fd6_37a1_42ad_b55f_10a6e4124f9d.slice/crio-9db3ef09001c6915793bec10648881df4667b0225691cb6ef6812c70de7cc0f4 WatchSource:0}: Error finding container 9db3ef09001c6915793bec10648881df4667b0225691cb6ef6812c70de7cc0f4: Status 404 returned error can't find the container with id 9db3ef09001c6915793bec10648881df4667b0225691cb6ef6812c70de7cc0f4 Dec 13 17:36:02 crc kubenswrapper[4989]: I1213 17:36:02.023296 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a999f01-db9b-4b04-9b64-1430f1e7eac1" path="/var/lib/kubelet/pods/0a999f01-db9b-4b04-9b64-1430f1e7eac1/volumes" Dec 13 17:36:02 crc kubenswrapper[4989]: I1213 17:36:02.024361 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c53a93e-56a6-49e5-9ee4-8113b9bb2693" path="/var/lib/kubelet/pods/5c53a93e-56a6-49e5-9ee4-8113b9bb2693/volumes" Dec 13 17:36:02 crc kubenswrapper[4989]: I1213 17:36:02.024862 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91505509-42c1-480e-9dbe-384203ec09b5" path="/var/lib/kubelet/pods/91505509-42c1-480e-9dbe-384203ec09b5/volumes" Dec 13 17:36:02 crc kubenswrapper[4989]: I1213 17:36:02.025893 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3419168-fb01-4821-b723-79aa7712cca4" path="/var/lib/kubelet/pods/d3419168-fb01-4821-b723-79aa7712cca4/volumes" Dec 13 17:36:02 crc kubenswrapper[4989]: I1213 17:36:02.265207 4989 generic.go:334] "Generic (PLEG): container finished" podID="27766e3d-62dd-4c50-85d5-e4ae824acd54" containerID="a44f8e7ed939b3d1b2466733ef84a4104089db0d54dbff2c85a0cbe753d51b83" exitCode=0 Dec 13 17:36:02 crc kubenswrapper[4989]: I1213 17:36:02.265293 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-78cjs" event={"ID":"27766e3d-62dd-4c50-85d5-e4ae824acd54","Type":"ContainerDied","Data":"a44f8e7ed939b3d1b2466733ef84a4104089db0d54dbff2c85a0cbe753d51b83"} Dec 13 17:36:02 crc kubenswrapper[4989]: I1213 17:36:02.265329 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-78cjs" event={"ID":"27766e3d-62dd-4c50-85d5-e4ae824acd54","Type":"ContainerStarted","Data":"43e9025a43ded95658769fb9e1fe782f972eab02b20190c7d49bfbdf5d81edf0"} Dec 13 17:36:02 crc kubenswrapper[4989]: I1213 17:36:02.266883 4989 generic.go:334] "Generic (PLEG): container finished" podID="dfb45fd6-37a1-42ad-b55f-10a6e4124f9d" containerID="d6ebdd145cc977673fa30d15692a37f80f5a420dd27941c18ed7d3f959d1a388" exitCode=0 Dec 13 17:36:02 crc kubenswrapper[4989]: I1213 17:36:02.266933 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" event={"ID":"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d","Type":"ContainerDied","Data":"d6ebdd145cc977673fa30d15692a37f80f5a420dd27941c18ed7d3f959d1a388"} Dec 13 17:36:02 crc kubenswrapper[4989]: I1213 17:36:02.266953 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" event={"ID":"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d","Type":"ContainerStarted","Data":"9db3ef09001c6915793bec10648881df4667b0225691cb6ef6812c70de7cc0f4"} Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.605997 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-78cjs" Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.695770 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.715911 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27766e3d-62dd-4c50-85d5-e4ae824acd54-operator-scripts\") pod \"27766e3d-62dd-4c50-85d5-e4ae824acd54\" (UID: \"27766e3d-62dd-4c50-85d5-e4ae824acd54\") " Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.716523 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz86j\" (UniqueName: \"kubernetes.io/projected/27766e3d-62dd-4c50-85d5-e4ae824acd54-kube-api-access-tz86j\") pod \"27766e3d-62dd-4c50-85d5-e4ae824acd54\" (UID: \"27766e3d-62dd-4c50-85d5-e4ae824acd54\") " Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.718425 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27766e3d-62dd-4c50-85d5-e4ae824acd54-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "27766e3d-62dd-4c50-85d5-e4ae824acd54" (UID: "27766e3d-62dd-4c50-85d5-e4ae824acd54"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.724806 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27766e3d-62dd-4c50-85d5-e4ae824acd54-kube-api-access-tz86j" (OuterVolumeSpecName: "kube-api-access-tz86j") pod "27766e3d-62dd-4c50-85d5-e4ae824acd54" (UID: "27766e3d-62dd-4c50-85d5-e4ae824acd54"). InnerVolumeSpecName "kube-api-access-tz86j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.817760 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-operator-scripts\") pod \"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d\" (UID: \"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d\") " Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.818027 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79wmv\" (UniqueName: \"kubernetes.io/projected/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-kube-api-access-79wmv\") pod \"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d\" (UID: \"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d\") " Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.818293 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dfb45fd6-37a1-42ad-b55f-10a6e4124f9d" (UID: "dfb45fd6-37a1-42ad-b55f-10a6e4124f9d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.818607 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.818634 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27766e3d-62dd-4c50-85d5-e4ae824acd54-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.818647 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz86j\" (UniqueName: \"kubernetes.io/projected/27766e3d-62dd-4c50-85d5-e4ae824acd54-kube-api-access-tz86j\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.820480 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-kube-api-access-79wmv" (OuterVolumeSpecName: "kube-api-access-79wmv") pod "dfb45fd6-37a1-42ad-b55f-10a6e4124f9d" (UID: "dfb45fd6-37a1-42ad-b55f-10a6e4124f9d"). InnerVolumeSpecName "kube-api-access-79wmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:03 crc kubenswrapper[4989]: I1213 17:36:03.919645 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79wmv\" (UniqueName: \"kubernetes.io/projected/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d-kube-api-access-79wmv\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:04 crc kubenswrapper[4989]: I1213 17:36:04.291845 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-78cjs" event={"ID":"27766e3d-62dd-4c50-85d5-e4ae824acd54","Type":"ContainerDied","Data":"43e9025a43ded95658769fb9e1fe782f972eab02b20190c7d49bfbdf5d81edf0"} Dec 13 17:36:04 crc kubenswrapper[4989]: I1213 17:36:04.292230 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-78cjs" Dec 13 17:36:04 crc kubenswrapper[4989]: I1213 17:36:04.292233 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43e9025a43ded95658769fb9e1fe782f972eab02b20190c7d49bfbdf5d81edf0" Dec 13 17:36:04 crc kubenswrapper[4989]: I1213 17:36:04.293764 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" event={"ID":"dfb45fd6-37a1-42ad-b55f-10a6e4124f9d","Type":"ContainerDied","Data":"9db3ef09001c6915793bec10648881df4667b0225691cb6ef6812c70de7cc0f4"} Dec 13 17:36:04 crc kubenswrapper[4989]: I1213 17:36:04.293816 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9db3ef09001c6915793bec10648881df4667b0225691cb6ef6812c70de7cc0f4" Dec 13 17:36:04 crc kubenswrapper[4989]: I1213 17:36:04.293861 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-8848-account-create-update-q84vz" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.149536 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-cmbl7"] Dec 13 17:36:06 crc kubenswrapper[4989]: E1213 17:36:06.149926 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb45fd6-37a1-42ad-b55f-10a6e4124f9d" containerName="mariadb-account-create-update" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.149950 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb45fd6-37a1-42ad-b55f-10a6e4124f9d" containerName="mariadb-account-create-update" Dec 13 17:36:06 crc kubenswrapper[4989]: E1213 17:36:06.149970 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27766e3d-62dd-4c50-85d5-e4ae824acd54" containerName="mariadb-database-create" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.149981 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="27766e3d-62dd-4c50-85d5-e4ae824acd54" containerName="mariadb-database-create" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.150196 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="27766e3d-62dd-4c50-85d5-e4ae824acd54" containerName="mariadb-database-create" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.150232 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfb45fd6-37a1-42ad-b55f-10a6e4124f9d" containerName="mariadb-account-create-update" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.150905 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.152951 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.153024 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.155330 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"combined-ca-bundle" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.155742 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-t456b" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.159684 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-cmbl7"] Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.253479 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-combined-ca-bundle\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.253541 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lzws\" (UniqueName: \"kubernetes.io/projected/f14134d9-da21-42b7-a38e-03a78a66b803-kube-api-access-4lzws\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.253584 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-db-sync-config-data\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.253625 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-config-data\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.253738 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-scripts\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.253802 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f14134d9-da21-42b7-a38e-03a78a66b803-etc-machine-id\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.354667 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-config-data\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.355507 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-scripts\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.355574 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f14134d9-da21-42b7-a38e-03a78a66b803-etc-machine-id\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.355597 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-combined-ca-bundle\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.355622 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lzws\" (UniqueName: \"kubernetes.io/projected/f14134d9-da21-42b7-a38e-03a78a66b803-kube-api-access-4lzws\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.355649 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-db-sync-config-data\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.356050 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f14134d9-da21-42b7-a38e-03a78a66b803-etc-machine-id\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.360660 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-db-sync-config-data\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.360835 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-config-data\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.366148 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-scripts\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.366187 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-combined-ca-bundle\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.378411 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lzws\" (UniqueName: \"kubernetes.io/projected/f14134d9-da21-42b7-a38e-03a78a66b803-kube-api-access-4lzws\") pod \"cinder-db-sync-cmbl7\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.467382 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:06 crc kubenswrapper[4989]: I1213 17:36:06.883378 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-cmbl7"] Dec 13 17:36:07 crc kubenswrapper[4989]: I1213 17:36:07.316008 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" event={"ID":"f14134d9-da21-42b7-a38e-03a78a66b803","Type":"ContainerStarted","Data":"708569f5b6d3a3719a4f9a23f9a2774e1d1805038c90dfd236a9e5e6ccdd7cea"} Dec 13 17:36:08 crc kubenswrapper[4989]: I1213 17:36:08.324632 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" event={"ID":"f14134d9-da21-42b7-a38e-03a78a66b803","Type":"ContainerStarted","Data":"ca86b6fae1da73c4c1ce6c36a602c4531e059363e24d326ec1d995e245f1a3ba"} Dec 13 17:36:08 crc kubenswrapper[4989]: I1213 17:36:08.344912 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" podStartSLOduration=2.344895073 podStartE2EDuration="2.344895073s" podCreationTimestamp="2025-12-13 17:36:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:36:08.340658772 +0000 UTC m=+1182.947105910" watchObservedRunningTime="2025-12-13 17:36:08.344895073 +0000 UTC m=+1182.951342211" Dec 13 17:36:10 crc kubenswrapper[4989]: I1213 17:36:10.341880 4989 generic.go:334] "Generic (PLEG): container finished" podID="f14134d9-da21-42b7-a38e-03a78a66b803" containerID="ca86b6fae1da73c4c1ce6c36a602c4531e059363e24d326ec1d995e245f1a3ba" exitCode=0 Dec 13 17:36:10 crc kubenswrapper[4989]: I1213 17:36:10.341994 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" event={"ID":"f14134d9-da21-42b7-a38e-03a78a66b803","Type":"ContainerDied","Data":"ca86b6fae1da73c4c1ce6c36a602c4531e059363e24d326ec1d995e245f1a3ba"} Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.659748 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.857078 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-scripts\") pod \"f14134d9-da21-42b7-a38e-03a78a66b803\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.857228 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-config-data\") pod \"f14134d9-da21-42b7-a38e-03a78a66b803\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.857274 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lzws\" (UniqueName: \"kubernetes.io/projected/f14134d9-da21-42b7-a38e-03a78a66b803-kube-api-access-4lzws\") pod \"f14134d9-da21-42b7-a38e-03a78a66b803\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.857317 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f14134d9-da21-42b7-a38e-03a78a66b803-etc-machine-id\") pod \"f14134d9-da21-42b7-a38e-03a78a66b803\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.857387 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-db-sync-config-data\") pod \"f14134d9-da21-42b7-a38e-03a78a66b803\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.857454 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-combined-ca-bundle\") pod \"f14134d9-da21-42b7-a38e-03a78a66b803\" (UID: \"f14134d9-da21-42b7-a38e-03a78a66b803\") " Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.858032 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f14134d9-da21-42b7-a38e-03a78a66b803-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f14134d9-da21-42b7-a38e-03a78a66b803" (UID: "f14134d9-da21-42b7-a38e-03a78a66b803"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.865154 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-scripts" (OuterVolumeSpecName: "scripts") pod "f14134d9-da21-42b7-a38e-03a78a66b803" (UID: "f14134d9-da21-42b7-a38e-03a78a66b803"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.865388 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f14134d9-da21-42b7-a38e-03a78a66b803" (UID: "f14134d9-da21-42b7-a38e-03a78a66b803"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.877881 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f14134d9-da21-42b7-a38e-03a78a66b803-kube-api-access-4lzws" (OuterVolumeSpecName: "kube-api-access-4lzws") pod "f14134d9-da21-42b7-a38e-03a78a66b803" (UID: "f14134d9-da21-42b7-a38e-03a78a66b803"). InnerVolumeSpecName "kube-api-access-4lzws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.883691 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f14134d9-da21-42b7-a38e-03a78a66b803" (UID: "f14134d9-da21-42b7-a38e-03a78a66b803"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.903487 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-config-data" (OuterVolumeSpecName: "config-data") pod "f14134d9-da21-42b7-a38e-03a78a66b803" (UID: "f14134d9-da21-42b7-a38e-03a78a66b803"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.960507 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.960564 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.960589 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lzws\" (UniqueName: \"kubernetes.io/projected/f14134d9-da21-42b7-a38e-03a78a66b803-kube-api-access-4lzws\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.960605 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f14134d9-da21-42b7-a38e-03a78a66b803-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.960620 4989 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:11 crc kubenswrapper[4989]: I1213 17:36:11.960633 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14134d9-da21-42b7-a38e-03a78a66b803-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.364798 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" event={"ID":"f14134d9-da21-42b7-a38e-03a78a66b803","Type":"ContainerDied","Data":"708569f5b6d3a3719a4f9a23f9a2774e1d1805038c90dfd236a9e5e6ccdd7cea"} Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.364853 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="708569f5b6d3a3719a4f9a23f9a2774e1d1805038c90dfd236a9e5e6ccdd7cea" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.364935 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-cmbl7" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.651486 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:36:12 crc kubenswrapper[4989]: E1213 17:36:12.652176 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f14134d9-da21-42b7-a38e-03a78a66b803" containerName="cinder-db-sync" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.652208 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f14134d9-da21-42b7-a38e-03a78a66b803" containerName="cinder-db-sync" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.652381 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f14134d9-da21-42b7-a38e-03a78a66b803" containerName="cinder-db-sync" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.656002 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.663232 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.663261 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"combined-ca-bundle" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.663736 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.665387 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scheduler-config-data" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.668191 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-t456b" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.670411 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.671551 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.674955 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-backup-config-data" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.696847 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.697937 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.700412 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-volume-volume1-config-data" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.719863 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.744880 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.770749 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.772627 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-sys\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.772666 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.772693 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-nvme\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.772725 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.772761 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.772803 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data-custom\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.772844 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.772907 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn6nz\" (UniqueName: \"kubernetes.io/projected/7a3c906a-a68f-4bed-8769-efa95ae69af5-kube-api-access-vn6nz\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.772962 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.772989 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7a3c906a-a68f-4bed-8769-efa95ae69af5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.773009 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.773037 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-lib-modules\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.773057 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.773078 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-dev\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.773101 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.773121 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-scripts\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.773141 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-run\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.773170 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.773188 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-scripts\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.773210 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvmrt\" (UniqueName: \"kubernetes.io/projected/46afd271-4300-4c6e-a046-befb2f11a6d8-kube-api-access-lvmrt\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.773231 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.874716 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7a3c906a-a68f-4bed-8769-efa95ae69af5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.875078 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.875187 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.875277 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-dev\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.875386 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-lib-modules\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.875477 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.875564 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.875653 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-dev\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.875741 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.875841 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-scripts\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.875982 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-run\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.876090 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-sys\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.876176 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.876253 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-scripts\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.876336 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvmrt\" (UniqueName: \"kubernetes.io/projected/46afd271-4300-4c6e-a046-befb2f11a6d8-kube-api-access-lvmrt\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.876952 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8g82\" (UniqueName: \"kubernetes.io/projected/67d04e42-14da-4e12-ad36-826e3bc3a6d3-kube-api-access-v8g82\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.877066 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.877154 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.876636 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-dev\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.876677 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.876546 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-lib-modules\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.875042 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7a3c906a-a68f-4bed-8769-efa95ae69af5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.877321 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.876597 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-run\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.877161 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.877698 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-sys\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.877798 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.877904 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.877992 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-nvme\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878090 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878162 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878190 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-run\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878236 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878291 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878364 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878396 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878471 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878528 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data-custom\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878556 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878605 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878635 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn6nz\" (UniqueName: \"kubernetes.io/projected/7a3c906a-a68f-4bed-8769-efa95ae69af5-kube-api-access-vn6nz\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878706 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878733 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878775 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878894 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-nvme\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878896 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.878920 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-sys\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.879663 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-scripts\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.879833 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-scripts\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.880332 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.884474 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.884892 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.897800 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data-custom\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.898874 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.902610 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.907265 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn6nz\" (UniqueName: \"kubernetes.io/projected/7a3c906a-a68f-4bed-8769-efa95ae69af5-kube-api-access-vn6nz\") pod \"cinder-scheduler-0\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.908346 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvmrt\" (UniqueName: \"kubernetes.io/projected/46afd271-4300-4c6e-a046-befb2f11a6d8-kube-api-access-lvmrt\") pod \"cinder-backup-0\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.935910 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.937005 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.940117 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cert-cinder-internal-svc" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.940575 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.943970 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cert-cinder-public-svc" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.948581 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980133 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980195 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-sys\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980221 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8g82\" (UniqueName: \"kubernetes.io/projected/67d04e42-14da-4e12-ad36-826e3bc3a6d3-kube-api-access-v8g82\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980244 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980265 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980296 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-run\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980312 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980332 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980357 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980382 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980402 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980432 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980454 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980480 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-dev\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.980503 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.981147 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-run\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.981214 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.981366 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.981409 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-sys\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.981439 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.981460 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.981478 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-dev\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.981511 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.981519 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.983062 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.984587 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.985367 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.986643 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:12 crc kubenswrapper[4989]: I1213 17:36:12.990131 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.000404 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.001011 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.001963 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8g82\" (UniqueName: \"kubernetes.io/projected/67d04e42-14da-4e12-ad36-826e3bc3a6d3-kube-api-access-v8g82\") pod \"cinder-volume-volume1-0\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.020772 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.082081 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.082142 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81d901a5-a913-4b33-89fa-40b82edf44e0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.082322 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.082470 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.082612 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81d901a5-a913-4b33-89fa-40b82edf44e0-logs\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.082711 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-scripts\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.082748 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54zhx\" (UniqueName: \"kubernetes.io/projected/81d901a5-a913-4b33-89fa-40b82edf44e0-kube-api-access-54zhx\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.082816 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.082933 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data-custom\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.184921 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-scripts\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.185333 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54zhx\" (UniqueName: \"kubernetes.io/projected/81d901a5-a913-4b33-89fa-40b82edf44e0-kube-api-access-54zhx\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.185361 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.185415 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data-custom\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.185454 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.185480 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81d901a5-a913-4b33-89fa-40b82edf44e0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.185511 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.185550 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.185610 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81d901a5-a913-4b33-89fa-40b82edf44e0-logs\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.186086 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81d901a5-a913-4b33-89fa-40b82edf44e0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.186159 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81d901a5-a913-4b33-89fa-40b82edf44e0-logs\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.191078 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-scripts\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.191947 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.192419 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data-custom\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.192691 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.193198 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.193984 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.202538 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54zhx\" (UniqueName: \"kubernetes.io/projected/81d901a5-a913-4b33-89fa-40b82edf44e0-kube-api-access-54zhx\") pod \"cinder-api-0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.308578 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.381056 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"7a3c906a-a68f-4bed-8769-efa95ae69af5","Type":"ContainerStarted","Data":"d4f2f76386c3d6dff8e4a64e1b10c008da28fbe34558fbe41caa1c9dcb25e895"} Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.390537 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.568888 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:36:13 crc kubenswrapper[4989]: W1213 17:36:13.584558 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67d04e42_14da_4e12_ad36_826e3bc3a6d3.slice/crio-2092c6fcbcf1a91316caa5f23a4c1c827f37c6d3cca059bf8e8efa141bba3a0b WatchSource:0}: Error finding container 2092c6fcbcf1a91316caa5f23a4c1c827f37c6d3cca059bf8e8efa141bba3a0b: Status 404 returned error can't find the container with id 2092c6fcbcf1a91316caa5f23a4c1c827f37c6d3cca059bf8e8efa141bba3a0b Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.587177 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:36:13 crc kubenswrapper[4989]: I1213 17:36:13.849115 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:36:14 crc kubenswrapper[4989]: I1213 17:36:14.398001 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"7a3c906a-a68f-4bed-8769-efa95ae69af5","Type":"ContainerStarted","Data":"3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c"} Dec 13 17:36:14 crc kubenswrapper[4989]: I1213 17:36:14.402849 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"81d901a5-a913-4b33-89fa-40b82edf44e0","Type":"ContainerStarted","Data":"4b3d1e508fb4f0943d3f85b61c2071993ed90df45767d48273f45953eee613aa"} Dec 13 17:36:14 crc kubenswrapper[4989]: I1213 17:36:14.405041 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"46afd271-4300-4c6e-a046-befb2f11a6d8","Type":"ContainerStarted","Data":"051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32"} Dec 13 17:36:14 crc kubenswrapper[4989]: I1213 17:36:14.405073 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"46afd271-4300-4c6e-a046-befb2f11a6d8","Type":"ContainerStarted","Data":"720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8"} Dec 13 17:36:14 crc kubenswrapper[4989]: I1213 17:36:14.405088 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"46afd271-4300-4c6e-a046-befb2f11a6d8","Type":"ContainerStarted","Data":"0004791c1b30ba1bc0d49f965a825fd5031da19ed6f14dacbb0844731cd94b08"} Dec 13 17:36:14 crc kubenswrapper[4989]: I1213 17:36:14.412699 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"67d04e42-14da-4e12-ad36-826e3bc3a6d3","Type":"ContainerStarted","Data":"cf3ec1e48f82daea8f99c74b029b49750ab11c9ad88a9e3d792c81af2487e62f"} Dec 13 17:36:14 crc kubenswrapper[4989]: I1213 17:36:14.412740 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"67d04e42-14da-4e12-ad36-826e3bc3a6d3","Type":"ContainerStarted","Data":"31b7aa8c78899e1b874832f28aa252ceb375b6a21c4fe00e32e40c3921c724e8"} Dec 13 17:36:14 crc kubenswrapper[4989]: I1213 17:36:14.412754 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"67d04e42-14da-4e12-ad36-826e3bc3a6d3","Type":"ContainerStarted","Data":"2092c6fcbcf1a91316caa5f23a4c1c827f37c6d3cca059bf8e8efa141bba3a0b"} Dec 13 17:36:14 crc kubenswrapper[4989]: I1213 17:36:14.435325 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-0" podStartSLOduration=2.435307873 podStartE2EDuration="2.435307873s" podCreationTimestamp="2025-12-13 17:36:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:36:14.430199984 +0000 UTC m=+1189.036647122" watchObservedRunningTime="2025-12-13 17:36:14.435307873 +0000 UTC m=+1189.041755011" Dec 13 17:36:14 crc kubenswrapper[4989]: I1213 17:36:14.459265 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podStartSLOduration=2.459247113 podStartE2EDuration="2.459247113s" podCreationTimestamp="2025-12-13 17:36:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:36:14.452151664 +0000 UTC m=+1189.058598812" watchObservedRunningTime="2025-12-13 17:36:14.459247113 +0000 UTC m=+1189.065694251" Dec 13 17:36:15 crc kubenswrapper[4989]: I1213 17:36:15.421350 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"81d901a5-a913-4b33-89fa-40b82edf44e0","Type":"ContainerStarted","Data":"4366ef7220c7a52e6151ef7ea90689c5a31be06cc79a7fa212785a4b85bc9c5a"} Dec 13 17:36:15 crc kubenswrapper[4989]: I1213 17:36:15.422033 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:15 crc kubenswrapper[4989]: I1213 17:36:15.422047 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"81d901a5-a913-4b33-89fa-40b82edf44e0","Type":"ContainerStarted","Data":"edd3d12c0cadc4f71e36d4b2100e11d999698286803e403c72af7832b6518fc7"} Dec 13 17:36:15 crc kubenswrapper[4989]: I1213 17:36:15.422997 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"7a3c906a-a68f-4bed-8769-efa95ae69af5","Type":"ContainerStarted","Data":"1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc"} Dec 13 17:36:15 crc kubenswrapper[4989]: I1213 17:36:15.438722 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=3.438701798 podStartE2EDuration="3.438701798s" podCreationTimestamp="2025-12-13 17:36:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:36:15.438347687 +0000 UTC m=+1190.044794825" watchObservedRunningTime="2025-12-13 17:36:15.438701798 +0000 UTC m=+1190.045148936" Dec 13 17:36:15 crc kubenswrapper[4989]: I1213 17:36:15.462739 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-0" podStartSLOduration=3.462721002 podStartE2EDuration="3.462721002s" podCreationTimestamp="2025-12-13 17:36:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:36:15.461071831 +0000 UTC m=+1190.067518969" watchObservedRunningTime="2025-12-13 17:36:15.462721002 +0000 UTC m=+1190.069168140" Dec 13 17:36:16 crc kubenswrapper[4989]: I1213 17:36:16.432595 4989 generic.go:334] "Generic (PLEG): container finished" podID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerID="cf3ec1e48f82daea8f99c74b029b49750ab11c9ad88a9e3d792c81af2487e62f" exitCode=1 Dec 13 17:36:16 crc kubenswrapper[4989]: I1213 17:36:16.433487 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"67d04e42-14da-4e12-ad36-826e3bc3a6d3","Type":"ContainerDied","Data":"cf3ec1e48f82daea8f99c74b029b49750ab11c9ad88a9e3d792c81af2487e62f"} Dec 13 17:36:16 crc kubenswrapper[4989]: I1213 17:36:16.433854 4989 scope.go:117] "RemoveContainer" containerID="cf3ec1e48f82daea8f99c74b029b49750ab11c9ad88a9e3d792c81af2487e62f" Dec 13 17:36:17 crc kubenswrapper[4989]: I1213 17:36:17.444929 4989 generic.go:334] "Generic (PLEG): container finished" podID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerID="31b7aa8c78899e1b874832f28aa252ceb375b6a21c4fe00e32e40c3921c724e8" exitCode=1 Dec 13 17:36:17 crc kubenswrapper[4989]: I1213 17:36:17.445276 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"67d04e42-14da-4e12-ad36-826e3bc3a6d3","Type":"ContainerDied","Data":"31b7aa8c78899e1b874832f28aa252ceb375b6a21c4fe00e32e40c3921c724e8"} Dec 13 17:36:17 crc kubenswrapper[4989]: I1213 17:36:17.445340 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"67d04e42-14da-4e12-ad36-826e3bc3a6d3","Type":"ContainerStarted","Data":"75a5a8db2992aa16119704b3de3c4c53812130853a0f2bfacbb25d6493247dfa"} Dec 13 17:36:17 crc kubenswrapper[4989]: I1213 17:36:17.445947 4989 scope.go:117] "RemoveContainer" containerID="31b7aa8c78899e1b874832f28aa252ceb375b6a21c4fe00e32e40c3921c724e8" Dec 13 17:36:17 crc kubenswrapper[4989]: I1213 17:36:17.991551 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:18 crc kubenswrapper[4989]: I1213 17:36:18.001165 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:18 crc kubenswrapper[4989]: I1213 17:36:18.026795 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:18 crc kubenswrapper[4989]: I1213 17:36:18.026924 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:18 crc kubenswrapper[4989]: I1213 17:36:18.251952 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:18 crc kubenswrapper[4989]: I1213 17:36:18.457815 4989 generic.go:334] "Generic (PLEG): container finished" podID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerID="75a5a8db2992aa16119704b3de3c4c53812130853a0f2bfacbb25d6493247dfa" exitCode=1 Dec 13 17:36:18 crc kubenswrapper[4989]: I1213 17:36:18.457933 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"67d04e42-14da-4e12-ad36-826e3bc3a6d3","Type":"ContainerDied","Data":"75a5a8db2992aa16119704b3de3c4c53812130853a0f2bfacbb25d6493247dfa"} Dec 13 17:36:18 crc kubenswrapper[4989]: I1213 17:36:18.458028 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"67d04e42-14da-4e12-ad36-826e3bc3a6d3","Type":"ContainerStarted","Data":"fffc2f2cec4f98a4869a1e863321ecc2a2050b7a1cb8aa580e288531991345b8"} Dec 13 17:36:18 crc kubenswrapper[4989]: I1213 17:36:18.458059 4989 scope.go:117] "RemoveContainer" containerID="cf3ec1e48f82daea8f99c74b029b49750ab11c9ad88a9e3d792c81af2487e62f" Dec 13 17:36:18 crc kubenswrapper[4989]: I1213 17:36:18.458491 4989 scope.go:117] "RemoveContainer" containerID="75a5a8db2992aa16119704b3de3c4c53812130853a0f2bfacbb25d6493247dfa" Dec 13 17:36:18 crc kubenswrapper[4989]: E1213 17:36:18.458748 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(67d04e42-14da-4e12-ad36-826e3bc3a6d3)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" Dec 13 17:36:19 crc kubenswrapper[4989]: I1213 17:36:19.467851 4989 generic.go:334] "Generic (PLEG): container finished" podID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerID="fffc2f2cec4f98a4869a1e863321ecc2a2050b7a1cb8aa580e288531991345b8" exitCode=1 Dec 13 17:36:19 crc kubenswrapper[4989]: I1213 17:36:19.467961 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"67d04e42-14da-4e12-ad36-826e3bc3a6d3","Type":"ContainerDied","Data":"fffc2f2cec4f98a4869a1e863321ecc2a2050b7a1cb8aa580e288531991345b8"} Dec 13 17:36:19 crc kubenswrapper[4989]: I1213 17:36:19.468365 4989 scope.go:117] "RemoveContainer" containerID="31b7aa8c78899e1b874832f28aa252ceb375b6a21c4fe00e32e40c3921c724e8" Dec 13 17:36:19 crc kubenswrapper[4989]: I1213 17:36:19.468588 4989 scope.go:117] "RemoveContainer" containerID="fffc2f2cec4f98a4869a1e863321ecc2a2050b7a1cb8aa580e288531991345b8" Dec 13 17:36:19 crc kubenswrapper[4989]: I1213 17:36:19.468657 4989 scope.go:117] "RemoveContainer" containerID="75a5a8db2992aa16119704b3de3c4c53812130853a0f2bfacbb25d6493247dfa" Dec 13 17:36:19 crc kubenswrapper[4989]: E1213 17:36:19.469035 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(67d04e42-14da-4e12-ad36-826e3bc3a6d3)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(67d04e42-14da-4e12-ad36-826e3bc3a6d3)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" Dec 13 17:36:20 crc kubenswrapper[4989]: I1213 17:36:20.480849 4989 scope.go:117] "RemoveContainer" containerID="fffc2f2cec4f98a4869a1e863321ecc2a2050b7a1cb8aa580e288531991345b8" Dec 13 17:36:20 crc kubenswrapper[4989]: I1213 17:36:20.480880 4989 scope.go:117] "RemoveContainer" containerID="75a5a8db2992aa16119704b3de3c4c53812130853a0f2bfacbb25d6493247dfa" Dec 13 17:36:20 crc kubenswrapper[4989]: E1213 17:36:20.481145 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(67d04e42-14da-4e12-ad36-826e3bc3a6d3)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(67d04e42-14da-4e12-ad36-826e3bc3a6d3)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" Dec 13 17:36:22 crc kubenswrapper[4989]: I1213 17:36:22.021564 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:22 crc kubenswrapper[4989]: I1213 17:36:22.022535 4989 scope.go:117] "RemoveContainer" containerID="fffc2f2cec4f98a4869a1e863321ecc2a2050b7a1cb8aa580e288531991345b8" Dec 13 17:36:22 crc kubenswrapper[4989]: I1213 17:36:22.022548 4989 scope.go:117] "RemoveContainer" containerID="75a5a8db2992aa16119704b3de3c4c53812130853a0f2bfacbb25d6493247dfa" Dec 13 17:36:22 crc kubenswrapper[4989]: E1213 17:36:22.022755 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(67d04e42-14da-4e12-ad36-826e3bc3a6d3)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(67d04e42-14da-4e12-ad36-826e3bc3a6d3)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" Dec 13 17:36:23 crc kubenswrapper[4989]: I1213 17:36:23.022140 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:23 crc kubenswrapper[4989]: I1213 17:36:23.022214 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:23 crc kubenswrapper[4989]: I1213 17:36:23.023158 4989 scope.go:117] "RemoveContainer" containerID="fffc2f2cec4f98a4869a1e863321ecc2a2050b7a1cb8aa580e288531991345b8" Dec 13 17:36:23 crc kubenswrapper[4989]: I1213 17:36:23.023191 4989 scope.go:117] "RemoveContainer" containerID="75a5a8db2992aa16119704b3de3c4c53812130853a0f2bfacbb25d6493247dfa" Dec 13 17:36:23 crc kubenswrapper[4989]: E1213 17:36:23.023652 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(67d04e42-14da-4e12-ad36-826e3bc3a6d3)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(67d04e42-14da-4e12-ad36-826e3bc3a6d3)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" Dec 13 17:36:23 crc kubenswrapper[4989]: I1213 17:36:23.218314 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.380114 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.843177 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-cmbl7"] Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.848046 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-cmbl7"] Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.863462 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.868428 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.868695 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="46afd271-4300-4c6e-a046-befb2f11a6d8" containerName="cinder-backup" containerID="cri-o://720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8" gracePeriod=30 Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.868856 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="46afd271-4300-4c6e-a046-befb2f11a6d8" containerName="probe" containerID="cri-o://051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32" gracePeriod=30 Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.873011 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.873240 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="7a3c906a-a68f-4bed-8769-efa95ae69af5" containerName="cinder-scheduler" containerID="cri-o://3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c" gracePeriod=30 Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.873347 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="7a3c906a-a68f-4bed-8769-efa95ae69af5" containerName="probe" containerID="cri-o://1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc" gracePeriod=30 Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.912290 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder8848-account-delete-6nlsk"] Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.913567 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.929380 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder8848-account-delete-6nlsk"] Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.937005 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.937449 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerName="cinder-api-log" containerID="cri-o://edd3d12c0cadc4f71e36d4b2100e11d999698286803e403c72af7832b6518fc7" gracePeriod=30 Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.937529 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerName="cinder-api" containerID="cri-o://4366ef7220c7a52e6151ef7ea90689c5a31be06cc79a7fa212785a4b85bc9c5a" gracePeriod=30 Dec 13 17:36:25 crc kubenswrapper[4989]: I1213 17:36:25.945678 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-0" podUID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.108:8776/healthcheck\": EOF" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.030615 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0ab2fff-73ca-4851-8712-2dec699e4f8f-operator-scripts\") pod \"cinder8848-account-delete-6nlsk\" (UID: \"c0ab2fff-73ca-4851-8712-2dec699e4f8f\") " pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.030692 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2gj7\" (UniqueName: \"kubernetes.io/projected/c0ab2fff-73ca-4851-8712-2dec699e4f8f-kube-api-access-c2gj7\") pod \"cinder8848-account-delete-6nlsk\" (UID: \"c0ab2fff-73ca-4851-8712-2dec699e4f8f\") " pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.031498 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f14134d9-da21-42b7-a38e-03a78a66b803" path="/var/lib/kubelet/pods/f14134d9-da21-42b7-a38e-03a78a66b803/volumes" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.131721 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0ab2fff-73ca-4851-8712-2dec699e4f8f-operator-scripts\") pod \"cinder8848-account-delete-6nlsk\" (UID: \"c0ab2fff-73ca-4851-8712-2dec699e4f8f\") " pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.131827 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2gj7\" (UniqueName: \"kubernetes.io/projected/c0ab2fff-73ca-4851-8712-2dec699e4f8f-kube-api-access-c2gj7\") pod \"cinder8848-account-delete-6nlsk\" (UID: \"c0ab2fff-73ca-4851-8712-2dec699e4f8f\") " pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.133117 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0ab2fff-73ca-4851-8712-2dec699e4f8f-operator-scripts\") pod \"cinder8848-account-delete-6nlsk\" (UID: \"c0ab2fff-73ca-4851-8712-2dec699e4f8f\") " pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.150335 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2gj7\" (UniqueName: \"kubernetes.io/projected/c0ab2fff-73ca-4851-8712-2dec699e4f8f-kube-api-access-c2gj7\") pod \"cinder8848-account-delete-6nlsk\" (UID: \"c0ab2fff-73ca-4851-8712-2dec699e4f8f\") " pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.243717 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.302258 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334586 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-brick\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334642 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-iscsi\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334663 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-lib-cinder\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334720 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-scripts\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334747 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334765 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-lib-modules\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334812 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8g82\" (UniqueName: \"kubernetes.io/projected/67d04e42-14da-4e12-ad36-826e3bc3a6d3-kube-api-access-v8g82\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334837 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-combined-ca-bundle\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334863 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data-custom\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334897 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-sys\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334934 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-nvme\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.334953 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-machine-id\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.335009 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-dev\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.335051 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-cinder\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.335083 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-run\") pod \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\" (UID: \"67d04e42-14da-4e12-ad36-826e3bc3a6d3\") " Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.335451 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-run" (OuterVolumeSpecName: "run") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.335491 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.370017 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.370077 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.370097 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.370610 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.374425 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-scripts" (OuterVolumeSpecName: "scripts") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.377752 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.377848 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-sys" (OuterVolumeSpecName: "sys") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.377882 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-dev" (OuterVolumeSpecName: "dev") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.377907 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.377938 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.378981 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d04e42-14da-4e12-ad36-826e3bc3a6d3-kube-api-access-v8g82" (OuterVolumeSpecName: "kube-api-access-v8g82") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "kube-api-access-v8g82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.419457 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.428124 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data" (OuterVolumeSpecName: "config-data") pod "67d04e42-14da-4e12-ad36-826e3bc3a6d3" (UID: "67d04e42-14da-4e12-ad36-826e3bc3a6d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436245 4989 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-dev\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436273 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436287 4989 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-run\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436295 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436304 4989 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436315 4989 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436324 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436332 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436340 4989 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436348 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8g82\" (UniqueName: \"kubernetes.io/projected/67d04e42-14da-4e12-ad36-826e3bc3a6d3-kube-api-access-v8g82\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436356 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436364 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67d04e42-14da-4e12-ad36-826e3bc3a6d3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436374 4989 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-sys\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436387 4989 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.436432 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67d04e42-14da-4e12-ad36-826e3bc3a6d3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.532821 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"67d04e42-14da-4e12-ad36-826e3bc3a6d3","Type":"ContainerDied","Data":"2092c6fcbcf1a91316caa5f23a4c1c827f37c6d3cca059bf8e8efa141bba3a0b"} Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.532873 4989 scope.go:117] "RemoveContainer" containerID="fffc2f2cec4f98a4869a1e863321ecc2a2050b7a1cb8aa580e288531991345b8" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.532978 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.553641 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"81d901a5-a913-4b33-89fa-40b82edf44e0","Type":"ContainerDied","Data":"edd3d12c0cadc4f71e36d4b2100e11d999698286803e403c72af7832b6518fc7"} Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.553585 4989 generic.go:334] "Generic (PLEG): container finished" podID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerID="edd3d12c0cadc4f71e36d4b2100e11d999698286803e403c72af7832b6518fc7" exitCode=143 Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.586434 4989 scope.go:117] "RemoveContainer" containerID="85ba483443ea227de3586c42b7900ddef8a75d1d6f060b12b7da7f5bf2fc13a5" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.588061 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.594506 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.607918 4989 scope.go:117] "RemoveContainer" containerID="75a5a8db2992aa16119704b3de3c4c53812130853a0f2bfacbb25d6493247dfa" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.618141 4989 scope.go:117] "RemoveContainer" containerID="fa9938b36b3e27a2d80644c38da7dd3d05ab6b023311f59fedd463bd18652212" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.661262 4989 scope.go:117] "RemoveContainer" containerID="e616b212936318cebfc0623337103dd737b227be59f40c43fda4c160f806f73a" Dec 13 17:36:26 crc kubenswrapper[4989]: I1213 17:36:26.717526 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder8848-account-delete-6nlsk"] Dec 13 17:36:26 crc kubenswrapper[4989]: W1213 17:36:26.737962 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0ab2fff_73ca_4851_8712_2dec699e4f8f.slice/crio-298f4665568fc75a0254d99bb8c300879e56de5af21c7a95e99b4d30572f305c WatchSource:0}: Error finding container 298f4665568fc75a0254d99bb8c300879e56de5af21c7a95e99b4d30572f305c: Status 404 returned error can't find the container with id 298f4665568fc75a0254d99bb8c300879e56de5af21c7a95e99b4d30572f305c Dec 13 17:36:27 crc kubenswrapper[4989]: I1213 17:36:27.563693 4989 generic.go:334] "Generic (PLEG): container finished" podID="46afd271-4300-4c6e-a046-befb2f11a6d8" containerID="051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32" exitCode=0 Dec 13 17:36:27 crc kubenswrapper[4989]: I1213 17:36:27.563755 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"46afd271-4300-4c6e-a046-befb2f11a6d8","Type":"ContainerDied","Data":"051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32"} Dec 13 17:36:27 crc kubenswrapper[4989]: I1213 17:36:27.566246 4989 generic.go:334] "Generic (PLEG): container finished" podID="c0ab2fff-73ca-4851-8712-2dec699e4f8f" containerID="4fbe5371891617db8be6bf086c4bc30a0c16287339133510b5da3348cb525e90" exitCode=0 Dec 13 17:36:27 crc kubenswrapper[4989]: I1213 17:36:27.566314 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" event={"ID":"c0ab2fff-73ca-4851-8712-2dec699e4f8f","Type":"ContainerDied","Data":"4fbe5371891617db8be6bf086c4bc30a0c16287339133510b5da3348cb525e90"} Dec 13 17:36:27 crc kubenswrapper[4989]: I1213 17:36:27.566334 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" event={"ID":"c0ab2fff-73ca-4851-8712-2dec699e4f8f","Type":"ContainerStarted","Data":"298f4665568fc75a0254d99bb8c300879e56de5af21c7a95e99b4d30572f305c"} Dec 13 17:36:27 crc kubenswrapper[4989]: I1213 17:36:27.568378 4989 generic.go:334] "Generic (PLEG): container finished" podID="7a3c906a-a68f-4bed-8769-efa95ae69af5" containerID="1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc" exitCode=0 Dec 13 17:36:27 crc kubenswrapper[4989]: I1213 17:36:27.568428 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"7a3c906a-a68f-4bed-8769-efa95ae69af5","Type":"ContainerDied","Data":"1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc"} Dec 13 17:36:28 crc kubenswrapper[4989]: I1213 17:36:28.023017 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" path="/var/lib/kubelet/pods/67d04e42-14da-4e12-ad36-826e3bc3a6d3/volumes" Dec 13 17:36:28 crc kubenswrapper[4989]: I1213 17:36:28.875161 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" Dec 13 17:36:28 crc kubenswrapper[4989]: I1213 17:36:28.976601 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2gj7\" (UniqueName: \"kubernetes.io/projected/c0ab2fff-73ca-4851-8712-2dec699e4f8f-kube-api-access-c2gj7\") pod \"c0ab2fff-73ca-4851-8712-2dec699e4f8f\" (UID: \"c0ab2fff-73ca-4851-8712-2dec699e4f8f\") " Dec 13 17:36:28 crc kubenswrapper[4989]: I1213 17:36:28.976653 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0ab2fff-73ca-4851-8712-2dec699e4f8f-operator-scripts\") pod \"c0ab2fff-73ca-4851-8712-2dec699e4f8f\" (UID: \"c0ab2fff-73ca-4851-8712-2dec699e4f8f\") " Dec 13 17:36:28 crc kubenswrapper[4989]: I1213 17:36:28.977545 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0ab2fff-73ca-4851-8712-2dec699e4f8f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c0ab2fff-73ca-4851-8712-2dec699e4f8f" (UID: "c0ab2fff-73ca-4851-8712-2dec699e4f8f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:28 crc kubenswrapper[4989]: I1213 17:36:28.984010 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ab2fff-73ca-4851-8712-2dec699e4f8f-kube-api-access-c2gj7" (OuterVolumeSpecName: "kube-api-access-c2gj7") pod "c0ab2fff-73ca-4851-8712-2dec699e4f8f" (UID: "c0ab2fff-73ca-4851-8712-2dec699e4f8f"). InnerVolumeSpecName "kube-api-access-c2gj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:29 crc kubenswrapper[4989]: I1213 17:36:29.078860 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2gj7\" (UniqueName: \"kubernetes.io/projected/c0ab2fff-73ca-4851-8712-2dec699e4f8f-kube-api-access-c2gj7\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:29 crc kubenswrapper[4989]: I1213 17:36:29.078904 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0ab2fff-73ca-4851-8712-2dec699e4f8f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:29 crc kubenswrapper[4989]: I1213 17:36:29.585975 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" event={"ID":"c0ab2fff-73ca-4851-8712-2dec699e4f8f","Type":"ContainerDied","Data":"298f4665568fc75a0254d99bb8c300879e56de5af21c7a95e99b4d30572f305c"} Dec 13 17:36:29 crc kubenswrapper[4989]: I1213 17:36:29.586016 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="298f4665568fc75a0254d99bb8c300879e56de5af21c7a95e99b4d30572f305c" Dec 13 17:36:29 crc kubenswrapper[4989]: I1213 17:36:29.586052 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder8848-account-delete-6nlsk" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.335067 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.108:8776/healthcheck\": read tcp 10.217.0.2:40586->10.217.0.108:8776: read: connection reset by peer" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.598621 4989 generic.go:334] "Generic (PLEG): container finished" podID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerID="4366ef7220c7a52e6151ef7ea90689c5a31be06cc79a7fa212785a4b85bc9c5a" exitCode=0 Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.598669 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"81d901a5-a913-4b33-89fa-40b82edf44e0","Type":"ContainerDied","Data":"4366ef7220c7a52e6151ef7ea90689c5a31be06cc79a7fa212785a4b85bc9c5a"} Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.663571 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.716479 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-internal-tls-certs\") pod \"81d901a5-a913-4b33-89fa-40b82edf44e0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.716587 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data\") pod \"81d901a5-a913-4b33-89fa-40b82edf44e0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.716614 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-scripts\") pod \"81d901a5-a913-4b33-89fa-40b82edf44e0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.716645 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54zhx\" (UniqueName: \"kubernetes.io/projected/81d901a5-a913-4b33-89fa-40b82edf44e0-kube-api-access-54zhx\") pod \"81d901a5-a913-4b33-89fa-40b82edf44e0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.716675 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-public-tls-certs\") pod \"81d901a5-a913-4b33-89fa-40b82edf44e0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.716706 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81d901a5-a913-4b33-89fa-40b82edf44e0-logs\") pod \"81d901a5-a913-4b33-89fa-40b82edf44e0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.716754 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81d901a5-a913-4b33-89fa-40b82edf44e0-etc-machine-id\") pod \"81d901a5-a913-4b33-89fa-40b82edf44e0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.716783 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data-custom\") pod \"81d901a5-a913-4b33-89fa-40b82edf44e0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.716827 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-combined-ca-bundle\") pod \"81d901a5-a913-4b33-89fa-40b82edf44e0\" (UID: \"81d901a5-a913-4b33-89fa-40b82edf44e0\") " Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.716919 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81d901a5-a913-4b33-89fa-40b82edf44e0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "81d901a5-a913-4b33-89fa-40b82edf44e0" (UID: "81d901a5-a913-4b33-89fa-40b82edf44e0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.717715 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81d901a5-a913-4b33-89fa-40b82edf44e0-logs" (OuterVolumeSpecName: "logs") pod "81d901a5-a913-4b33-89fa-40b82edf44e0" (UID: "81d901a5-a913-4b33-89fa-40b82edf44e0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.718100 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/81d901a5-a913-4b33-89fa-40b82edf44e0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.718121 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81d901a5-a913-4b33-89fa-40b82edf44e0-logs\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.723184 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-scripts" (OuterVolumeSpecName: "scripts") pod "81d901a5-a913-4b33-89fa-40b82edf44e0" (UID: "81d901a5-a913-4b33-89fa-40b82edf44e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.724173 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81d901a5-a913-4b33-89fa-40b82edf44e0-kube-api-access-54zhx" (OuterVolumeSpecName: "kube-api-access-54zhx") pod "81d901a5-a913-4b33-89fa-40b82edf44e0" (UID: "81d901a5-a913-4b33-89fa-40b82edf44e0"). InnerVolumeSpecName "kube-api-access-54zhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.725418 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "81d901a5-a913-4b33-89fa-40b82edf44e0" (UID: "81d901a5-a913-4b33-89fa-40b82edf44e0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.736476 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81d901a5-a913-4b33-89fa-40b82edf44e0" (UID: "81d901a5-a913-4b33-89fa-40b82edf44e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.751353 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data" (OuterVolumeSpecName: "config-data") pod "81d901a5-a913-4b33-89fa-40b82edf44e0" (UID: "81d901a5-a913-4b33-89fa-40b82edf44e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.752921 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "81d901a5-a913-4b33-89fa-40b82edf44e0" (UID: "81d901a5-a913-4b33-89fa-40b82edf44e0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.753376 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "81d901a5-a913-4b33-89fa-40b82edf44e0" (UID: "81d901a5-a913-4b33-89fa-40b82edf44e0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.819957 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.820238 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.820257 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.820279 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.820297 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.820313 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54zhx\" (UniqueName: \"kubernetes.io/projected/81d901a5-a913-4b33-89fa-40b82edf44e0-kube-api-access-54zhx\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.820330 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d901a5-a913-4b33-89fa-40b82edf44e0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.947692 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-78cjs"] Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.955120 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-78cjs"] Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.962221 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-8848-account-create-update-q84vz"] Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.968600 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder8848-account-delete-6nlsk"] Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.974751 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-8848-account-create-update-q84vz"] Dec 13 17:36:30 crc kubenswrapper[4989]: I1213 17:36:30.981810 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder8848-account-delete-6nlsk"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.209640 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.266921 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326129 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-brick\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326180 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-sys\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326216 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvmrt\" (UniqueName: \"kubernetes.io/projected/46afd271-4300-4c6e-a046-befb2f11a6d8-kube-api-access-lvmrt\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326235 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-dev\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326249 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-cinder\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326267 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-combined-ca-bundle\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326268 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326300 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7a3c906a-a68f-4bed-8769-efa95ae69af5-etc-machine-id\") pod \"7a3c906a-a68f-4bed-8769-efa95ae69af5\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326268 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-sys" (OuterVolumeSpecName: "sys") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326326 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-scripts\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326354 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-scripts\") pod \"7a3c906a-a68f-4bed-8769-efa95ae69af5\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326366 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-dev" (OuterVolumeSpecName: "dev") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326381 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-machine-id\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326399 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326431 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326406 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a3c906a-a68f-4bed-8769-efa95ae69af5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7a3c906a-a68f-4bed-8769-efa95ae69af5" (UID: "7a3c906a-a68f-4bed-8769-efa95ae69af5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326455 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-run\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326478 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-run" (OuterVolumeSpecName: "run") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326490 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data-custom\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326516 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn6nz\" (UniqueName: \"kubernetes.io/projected/7a3c906a-a68f-4bed-8769-efa95ae69af5-kube-api-access-vn6nz\") pod \"7a3c906a-a68f-4bed-8769-efa95ae69af5\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326549 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data-custom\") pod \"7a3c906a-a68f-4bed-8769-efa95ae69af5\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326569 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326595 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-nvme\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326617 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data\") pod \"7a3c906a-a68f-4bed-8769-efa95ae69af5\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326641 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-lib-cinder\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326676 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-iscsi\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326688 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326702 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-lib-modules\") pod \"46afd271-4300-4c6e-a046-befb2f11a6d8\" (UID: \"46afd271-4300-4c6e-a046-befb2f11a6d8\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326742 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326748 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-combined-ca-bundle\") pod \"7a3c906a-a68f-4bed-8769-efa95ae69af5\" (UID: \"7a3c906a-a68f-4bed-8769-efa95ae69af5\") " Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326814 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.326868 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.327436 4989 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.327455 4989 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.327467 4989 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.327478 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.327489 4989 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-sys\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.327500 4989 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-dev\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.327510 4989 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.327520 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7a3c906a-a68f-4bed-8769-efa95ae69af5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.327531 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.327541 4989 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-run\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.327551 4989 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/46afd271-4300-4c6e-a046-befb2f11a6d8-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.330431 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.330446 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7a3c906a-a68f-4bed-8769-efa95ae69af5" (UID: "7a3c906a-a68f-4bed-8769-efa95ae69af5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.330715 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46afd271-4300-4c6e-a046-befb2f11a6d8-kube-api-access-lvmrt" (OuterVolumeSpecName: "kube-api-access-lvmrt") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "kube-api-access-lvmrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.330898 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a3c906a-a68f-4bed-8769-efa95ae69af5-kube-api-access-vn6nz" (OuterVolumeSpecName: "kube-api-access-vn6nz") pod "7a3c906a-a68f-4bed-8769-efa95ae69af5" (UID: "7a3c906a-a68f-4bed-8769-efa95ae69af5"). InnerVolumeSpecName "kube-api-access-vn6nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.331332 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-scripts" (OuterVolumeSpecName: "scripts") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.331885 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-scripts" (OuterVolumeSpecName: "scripts") pod "7a3c906a-a68f-4bed-8769-efa95ae69af5" (UID: "7a3c906a-a68f-4bed-8769-efa95ae69af5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.357159 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a3c906a-a68f-4bed-8769-efa95ae69af5" (UID: "7a3c906a-a68f-4bed-8769-efa95ae69af5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.359089 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.380563 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data" (OuterVolumeSpecName: "config-data") pod "7a3c906a-a68f-4bed-8769-efa95ae69af5" (UID: "7a3c906a-a68f-4bed-8769-efa95ae69af5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.394682 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data" (OuterVolumeSpecName: "config-data") pod "46afd271-4300-4c6e-a046-befb2f11a6d8" (UID: "46afd271-4300-4c6e-a046-befb2f11a6d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.428573 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.428607 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.428616 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.428624 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.428635 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvmrt\" (UniqueName: \"kubernetes.io/projected/46afd271-4300-4c6e-a046-befb2f11a6d8-kube-api-access-lvmrt\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.428644 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.428652 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.428660 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a3c906a-a68f-4bed-8769-efa95ae69af5-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.428668 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46afd271-4300-4c6e-a046-befb2f11a6d8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.428676 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn6nz\" (UniqueName: \"kubernetes.io/projected/7a3c906a-a68f-4bed-8769-efa95ae69af5-kube-api-access-vn6nz\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.609488 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.609550 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"7a3c906a-a68f-4bed-8769-efa95ae69af5","Type":"ContainerDied","Data":"3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c"} Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.609599 4989 scope.go:117] "RemoveContainer" containerID="1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.609394 4989 generic.go:334] "Generic (PLEG): container finished" podID="7a3c906a-a68f-4bed-8769-efa95ae69af5" containerID="3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c" exitCode=0 Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.613269 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"7a3c906a-a68f-4bed-8769-efa95ae69af5","Type":"ContainerDied","Data":"d4f2f76386c3d6dff8e4a64e1b10c008da28fbe34558fbe41caa1c9dcb25e895"} Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.615777 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"81d901a5-a913-4b33-89fa-40b82edf44e0","Type":"ContainerDied","Data":"4b3d1e508fb4f0943d3f85b61c2071993ed90df45767d48273f45953eee613aa"} Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.615875 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.618368 4989 generic.go:334] "Generic (PLEG): container finished" podID="46afd271-4300-4c6e-a046-befb2f11a6d8" containerID="720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8" exitCode=0 Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.618411 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"46afd271-4300-4c6e-a046-befb2f11a6d8","Type":"ContainerDied","Data":"720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8"} Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.618425 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.618437 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"46afd271-4300-4c6e-a046-befb2f11a6d8","Type":"ContainerDied","Data":"0004791c1b30ba1bc0d49f965a825fd5031da19ed6f14dacbb0844731cd94b08"} Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.631547 4989 scope.go:117] "RemoveContainer" containerID="3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.640859 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.646592 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.656338 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.657978 4989 scope.go:117] "RemoveContainer" containerID="1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.659588 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc\": container with ID starting with 1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc not found: ID does not exist" containerID="1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.659627 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc"} err="failed to get container status \"1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc\": rpc error: code = NotFound desc = could not find container \"1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc\": container with ID starting with 1978cd8586d9f1c08e235bb0af02dfafab0daa026d572c8e8b04024718782afc not found: ID does not exist" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.659648 4989 scope.go:117] "RemoveContainer" containerID="3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.660285 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c\": container with ID starting with 3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c not found: ID does not exist" containerID="3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.660311 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c"} err="failed to get container status \"3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c\": rpc error: code = NotFound desc = could not find container \"3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c\": container with ID starting with 3fa1d11a6aa2cb606536d61c5de9677c84e3809c008fe5761b8f980b2f6bd11c not found: ID does not exist" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.660348 4989 scope.go:117] "RemoveContainer" containerID="4366ef7220c7a52e6151ef7ea90689c5a31be06cc79a7fa212785a4b85bc9c5a" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.661679 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.672583 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.677461 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.688743 4989 scope.go:117] "RemoveContainer" containerID="edd3d12c0cadc4f71e36d4b2100e11d999698286803e403c72af7832b6518fc7" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.713037 4989 scope.go:117] "RemoveContainer" containerID="051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.733767 4989 scope.go:117] "RemoveContainer" containerID="720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.764609 4989 scope.go:117] "RemoveContainer" containerID="051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.765116 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32\": container with ID starting with 051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32 not found: ID does not exist" containerID="051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.765155 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32"} err="failed to get container status \"051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32\": rpc error: code = NotFound desc = could not find container \"051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32\": container with ID starting with 051cc887eb9eaa725881bc49a57eabbf568896b3fb173c0eb37f6f392718be32 not found: ID does not exist" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.765187 4989 scope.go:117] "RemoveContainer" containerID="720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.765695 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8\": container with ID starting with 720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8 not found: ID does not exist" containerID="720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.765769 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8"} err="failed to get container status \"720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8\": rpc error: code = NotFound desc = could not find container \"720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8\": container with ID starting with 720c6ca81b4c5efad4e7f8457eeb1c42aa318e7ee00982ab57a0b4b76ebb46b8 not found: ID does not exist" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.771294 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-66jpr"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.778004 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-66jpr"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.783519 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-7wrn9"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.790319 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-7wrn9"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.814299 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-58668b4874-gcpk9"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.814582 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" podUID="69501fce-9a39-435d-a689-6b1eddcf7188" containerName="keystone-api" containerID="cri-o://64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852" gracePeriod=30 Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.824115 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone7987-account-delete-qq7zw"] Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.825604 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="probe" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.825639 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="probe" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.825655 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a3c906a-a68f-4bed-8769-efa95ae69af5" containerName="probe" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.825662 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a3c906a-a68f-4bed-8769-efa95ae69af5" containerName="probe" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.825691 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ab2fff-73ca-4851-8712-2dec699e4f8f" containerName="mariadb-account-delete" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.825698 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ab2fff-73ca-4851-8712-2dec699e4f8f" containerName="mariadb-account-delete" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.825714 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a3c906a-a68f-4bed-8769-efa95ae69af5" containerName="cinder-scheduler" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.825724 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a3c906a-a68f-4bed-8769-efa95ae69af5" containerName="cinder-scheduler" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.825740 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46afd271-4300-4c6e-a046-befb2f11a6d8" containerName="cinder-backup" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.825747 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="46afd271-4300-4c6e-a046-befb2f11a6d8" containerName="cinder-backup" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.825755 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46afd271-4300-4c6e-a046-befb2f11a6d8" containerName="probe" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.825762 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="46afd271-4300-4c6e-a046-befb2f11a6d8" containerName="probe" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.825777 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerName="cinder-api-log" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.825787 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerName="cinder-api-log" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.825819 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="cinder-volume" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.825826 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="cinder-volume" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.825837 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerName="cinder-api" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.825845 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerName="cinder-api" Dec 13 17:36:31 crc kubenswrapper[4989]: E1213 17:36:31.825860 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="cinder-volume" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.825867 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="cinder-volume" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.826450 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a3c906a-a68f-4bed-8769-efa95ae69af5" containerName="probe" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.826463 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="46afd271-4300-4c6e-a046-befb2f11a6d8" containerName="probe" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.826487 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a3c906a-a68f-4bed-8769-efa95ae69af5" containerName="cinder-scheduler" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.826495 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="cinder-volume" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.826508 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="cinder-volume" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.826521 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="probe" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.826533 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="probe" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.826546 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="46afd271-4300-4c6e-a046-befb2f11a6d8" containerName="cinder-backup" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.826563 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerName="cinder-api" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.826596 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ab2fff-73ca-4851-8712-2dec699e4f8f" containerName="mariadb-account-delete" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.826610 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="81d901a5-a913-4b33-89fa-40b82edf44e0" containerName="cinder-api-log" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.827697 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.833415 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone7987-account-delete-qq7zw"] Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.953289 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b8tp\" (UniqueName: \"kubernetes.io/projected/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-kube-api-access-2b8tp\") pod \"keystone7987-account-delete-qq7zw\" (UID: \"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216\") " pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" Dec 13 17:36:31 crc kubenswrapper[4989]: I1213 17:36:31.953345 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts\") pod \"keystone7987-account-delete-qq7zw\" (UID: \"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216\") " pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.023136 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27766e3d-62dd-4c50-85d5-e4ae824acd54" path="/var/lib/kubelet/pods/27766e3d-62dd-4c50-85d5-e4ae824acd54/volumes" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.023778 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46afd271-4300-4c6e-a046-befb2f11a6d8" path="/var/lib/kubelet/pods/46afd271-4300-4c6e-a046-befb2f11a6d8/volumes" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.024448 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a3c906a-a68f-4bed-8769-efa95ae69af5" path="/var/lib/kubelet/pods/7a3c906a-a68f-4bed-8769-efa95ae69af5/volumes" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.025578 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81d901a5-a913-4b33-89fa-40b82edf44e0" path="/var/lib/kubelet/pods/81d901a5-a913-4b33-89fa-40b82edf44e0/volumes" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.026144 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0587684-6ad8-4723-8fcc-bcb8abbbae22" path="/var/lib/kubelet/pods/c0587684-6ad8-4723-8fcc-bcb8abbbae22/volumes" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.026657 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0ab2fff-73ca-4851-8712-2dec699e4f8f" path="/var/lib/kubelet/pods/c0ab2fff-73ca-4851-8712-2dec699e4f8f/volumes" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.027666 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfb45fd6-37a1-42ad-b55f-10a6e4124f9d" path="/var/lib/kubelet/pods/dfb45fd6-37a1-42ad-b55f-10a6e4124f9d/volumes" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.028165 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e61948e3-aa05-40c9-92be-ff2eb7bc9412" path="/var/lib/kubelet/pods/e61948e3-aa05-40c9-92be-ff2eb7bc9412/volumes" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.054953 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b8tp\" (UniqueName: \"kubernetes.io/projected/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-kube-api-access-2b8tp\") pod \"keystone7987-account-delete-qq7zw\" (UID: \"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216\") " pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.055004 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts\") pod \"keystone7987-account-delete-qq7zw\" (UID: \"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216\") " pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.055728 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts\") pod \"keystone7987-account-delete-qq7zw\" (UID: \"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216\") " pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.076834 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b8tp\" (UniqueName: \"kubernetes.io/projected/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-kube-api-access-2b8tp\") pod \"keystone7987-account-delete-qq7zw\" (UID: \"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216\") " pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" Dec 13 17:36:32 crc kubenswrapper[4989]: I1213 17:36:32.178080 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:32.419049 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-hllmp"] Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:32.431182 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-hllmp"] Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:32.449894 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:32.456479 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:32.463826 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:32.587947 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-2" podUID="c540cd7f-4b50-4a71-9e51-0687156ee1b2" containerName="galera" containerID="cri-o://28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c" gracePeriod=30 Dec 13 17:36:33 crc kubenswrapper[4989]: E1213 17:36:32.900218 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 13 17:36:33 crc kubenswrapper[4989]: E1213 17:36:32.901536 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 13 17:36:33 crc kubenswrapper[4989]: E1213 17:36:32.902900 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 13 17:36:33 crc kubenswrapper[4989]: E1213 17:36:32.902935 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="cinder-kuttl-tests/openstack-galera-2" podUID="c540cd7f-4b50-4a71-9e51-0687156ee1b2" containerName="galera" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.206368 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.207109 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/memcached-0" podUID="bb43b605-721a-4dae-98c8-eeabd32ba579" containerName="memcached" containerID="cri-o://53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf" gracePeriod=30 Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.385377 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone7987-account-delete-qq7zw"] Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.491707 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.577547 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-default\") pod \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.577630 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kolla-config\") pod \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.577699 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.577728 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-operator-scripts\") pod \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.577828 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-generated\") pod \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.577877 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95t6f\" (UniqueName: \"kubernetes.io/projected/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kube-api-access-95t6f\") pod \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\" (UID: \"c540cd7f-4b50-4a71-9e51-0687156ee1b2\") " Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.578220 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "c540cd7f-4b50-4a71-9e51-0687156ee1b2" (UID: "c540cd7f-4b50-4a71-9e51-0687156ee1b2"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.578308 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "c540cd7f-4b50-4a71-9e51-0687156ee1b2" (UID: "c540cd7f-4b50-4a71-9e51-0687156ee1b2"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.578351 4989 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.579502 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "c540cd7f-4b50-4a71-9e51-0687156ee1b2" (UID: "c540cd7f-4b50-4a71-9e51-0687156ee1b2"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.579707 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c540cd7f-4b50-4a71-9e51-0687156ee1b2" (UID: "c540cd7f-4b50-4a71-9e51-0687156ee1b2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.592055 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kube-api-access-95t6f" (OuterVolumeSpecName: "kube-api-access-95t6f") pod "c540cd7f-4b50-4a71-9e51-0687156ee1b2" (UID: "c540cd7f-4b50-4a71-9e51-0687156ee1b2"). InnerVolumeSpecName "kube-api-access-95t6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.593205 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "c540cd7f-4b50-4a71-9e51-0687156ee1b2" (UID: "c540cd7f-4b50-4a71-9e51-0687156ee1b2"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.650004 4989 generic.go:334] "Generic (PLEG): container finished" podID="c540cd7f-4b50-4a71-9e51-0687156ee1b2" containerID="28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c" exitCode=0 Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.650080 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"c540cd7f-4b50-4a71-9e51-0687156ee1b2","Type":"ContainerDied","Data":"28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c"} Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.650114 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"c540cd7f-4b50-4a71-9e51-0687156ee1b2","Type":"ContainerDied","Data":"1c866768a930d535c9685f7938676ec6e51e35edb7ff02f9c5bf1d628f6b75ff"} Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.650133 4989 scope.go:117] "RemoveContainer" containerID="28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.650288 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.653225 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" event={"ID":"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216","Type":"ContainerStarted","Data":"c676121bddd5b9c87e39c6c8a793e13114bf825bde93b16b72a4b92d2738456a"} Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.653306 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" event={"ID":"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216","Type":"ContainerStarted","Data":"120531dbd988e3ec454a4b8fbf402f2c247c7edaa2f3271e7526a055622e3e15"} Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.654011 4989 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" secret="" err="secret \"galera-openstack-dockercfg-gvwv4\" not found" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.678539 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" podStartSLOduration=2.678514389 podStartE2EDuration="2.678514389s" podCreationTimestamp="2025-12-13 17:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:36:33.668684694 +0000 UTC m=+1208.275131832" watchObservedRunningTime="2025-12-13 17:36:33.678514389 +0000 UTC m=+1208.284961537" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.679452 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.679480 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95t6f\" (UniqueName: \"kubernetes.io/projected/c540cd7f-4b50-4a71-9e51-0687156ee1b2-kube-api-access-95t6f\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.679493 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.679528 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.679542 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c540cd7f-4b50-4a71-9e51-0687156ee1b2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.694043 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.696656 4989 scope.go:117] "RemoveContainer" containerID="df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.701407 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.701614 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.722730 4989 scope.go:117] "RemoveContainer" containerID="28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c" Dec 13 17:36:33 crc kubenswrapper[4989]: E1213 17:36:33.723245 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c\": container with ID starting with 28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c not found: ID does not exist" containerID="28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.723292 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c"} err="failed to get container status \"28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c\": rpc error: code = NotFound desc = could not find container \"28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c\": container with ID starting with 28c32ae8801a3297d076139b77196279157ddb71ad6226e17bf9673df83b682c not found: ID does not exist" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.723320 4989 scope.go:117] "RemoveContainer" containerID="df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6" Dec 13 17:36:33 crc kubenswrapper[4989]: E1213 17:36:33.723852 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6\": container with ID starting with df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6 not found: ID does not exist" containerID="df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.723882 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6"} err="failed to get container status \"df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6\": rpc error: code = NotFound desc = could not find container \"df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6\": container with ID starting with df9b03ee0db1ee14cfa58ce20dbbb4bee08228ee3ed4d2f04b7c6e15bdc1f6d6 not found: ID does not exist" Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.781848 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:33 crc kubenswrapper[4989]: E1213 17:36:33.781965 4989 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Dec 13 17:36:33 crc kubenswrapper[4989]: E1213 17:36:33.782042 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts podName:7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216 nodeName:}" failed. No retries permitted until 2025-12-13 17:36:34.282018224 +0000 UTC m=+1208.888465372 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts") pod "keystone7987-account-delete-qq7zw" (UID: "7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216") : configmap "openstack-scripts" not found Dec 13 17:36:33 crc kubenswrapper[4989]: I1213 17:36:33.800033 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.023835 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74cceb6b-b34b-42c2-8fe4-b93433e758d0" path="/var/lib/kubelet/pods/74cceb6b-b34b-42c2-8fe4-b93433e758d0/volumes" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.025591 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c540cd7f-4b50-4a71-9e51-0687156ee1b2" path="/var/lib/kubelet/pods/c540cd7f-4b50-4a71-9e51-0687156ee1b2/volumes" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.199329 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 13 17:36:34 crc kubenswrapper[4989]: E1213 17:36:34.294899 4989 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Dec 13 17:36:34 crc kubenswrapper[4989]: E1213 17:36:34.294964 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts podName:7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216 nodeName:}" failed. No retries permitted until 2025-12-13 17:36:35.294949195 +0000 UTC m=+1209.901396333 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts") pod "keystone7987-account-delete-qq7zw" (UID: "7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216") : configmap "openstack-scripts" not found Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.514123 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.600096 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-config-data\") pod \"bb43b605-721a-4dae-98c8-eeabd32ba579\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.600206 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rml4\" (UniqueName: \"kubernetes.io/projected/bb43b605-721a-4dae-98c8-eeabd32ba579-kube-api-access-8rml4\") pod \"bb43b605-721a-4dae-98c8-eeabd32ba579\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.600248 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-kolla-config\") pod \"bb43b605-721a-4dae-98c8-eeabd32ba579\" (UID: \"bb43b605-721a-4dae-98c8-eeabd32ba579\") " Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.600971 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-config-data" (OuterVolumeSpecName: "config-data") pod "bb43b605-721a-4dae-98c8-eeabd32ba579" (UID: "bb43b605-721a-4dae-98c8-eeabd32ba579"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.601178 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "bb43b605-721a-4dae-98c8-eeabd32ba579" (UID: "bb43b605-721a-4dae-98c8-eeabd32ba579"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.602878 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-1" podUID="b50061aa-30a3-4145-bca8-94c802594dc4" containerName="galera" containerID="cri-o://d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c" gracePeriod=28 Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.606320 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb43b605-721a-4dae-98c8-eeabd32ba579-kube-api-access-8rml4" (OuterVolumeSpecName: "kube-api-access-8rml4") pod "bb43b605-721a-4dae-98c8-eeabd32ba579" (UID: "bb43b605-721a-4dae-98c8-eeabd32ba579"). InnerVolumeSpecName "kube-api-access-8rml4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.663983 4989 generic.go:334] "Generic (PLEG): container finished" podID="7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" containerID="c676121bddd5b9c87e39c6c8a793e13114bf825bde93b16b72a4b92d2738456a" exitCode=1 Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.664032 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" event={"ID":"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216","Type":"ContainerDied","Data":"c676121bddd5b9c87e39c6c8a793e13114bf825bde93b16b72a4b92d2738456a"} Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.665300 4989 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" secret="" err="secret \"galera-openstack-dockercfg-gvwv4\" not found" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.665385 4989 generic.go:334] "Generic (PLEG): container finished" podID="bb43b605-721a-4dae-98c8-eeabd32ba579" containerID="53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf" exitCode=0 Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.665431 4989 scope.go:117] "RemoveContainer" containerID="c676121bddd5b9c87e39c6c8a793e13114bf825bde93b16b72a4b92d2738456a" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.665465 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.665503 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"bb43b605-721a-4dae-98c8-eeabd32ba579","Type":"ContainerDied","Data":"53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf"} Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.665522 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"bb43b605-721a-4dae-98c8-eeabd32ba579","Type":"ContainerDied","Data":"cce1c659efffb47c75f85a472e8c0325c4434956e81648b29411f03894bce5a3"} Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.665539 4989 scope.go:117] "RemoveContainer" containerID="53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.702509 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.702541 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rml4\" (UniqueName: \"kubernetes.io/projected/bb43b605-721a-4dae-98c8-eeabd32ba579-kube-api-access-8rml4\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.702552 4989 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bb43b605-721a-4dae-98c8-eeabd32ba579-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.706132 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.707388 4989 scope.go:117] "RemoveContainer" containerID="53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf" Dec 13 17:36:34 crc kubenswrapper[4989]: E1213 17:36:34.707890 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf\": container with ID starting with 53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf not found: ID does not exist" containerID="53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.707917 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf"} err="failed to get container status \"53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf\": rpc error: code = NotFound desc = could not find container \"53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf\": container with ID starting with 53e6ab78b6d4461ccc267cf25894cda4569745f7b7904fbad9719072c459c4bf not found: ID does not exist" Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.711856 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 13 17:36:34 crc kubenswrapper[4989]: I1213 17:36:34.718053 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/rabbitmq-server-0" podUID="78da11d7-96b7-4cfd-b190-f34622427b21" containerName="rabbitmq" containerID="cri-o://9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb" gracePeriod=604800 Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.057269 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv"] Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.057950 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" podUID="93cc60c5-709e-474e-81c5-04242bbd0dec" containerName="manager" containerID="cri-o://aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482" gracePeriod=10 Dec 13 17:36:35 crc kubenswrapper[4989]: E1213 17:36:35.310130 4989 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Dec 13 17:36:35 crc kubenswrapper[4989]: E1213 17:36:35.310217 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts podName:7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216 nodeName:}" failed. No retries permitted until 2025-12-13 17:36:37.310195719 +0000 UTC m=+1211.916642857 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts") pod "keystone7987-account-delete-qq7zw" (UID: "7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216") : configmap "openstack-scripts" not found Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.342661 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-index-zjpg6"] Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.346972 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/cinder-operator-index-zjpg6" podUID="f814e529-71ee-4f12-8057-ebfe2cd78b97" containerName="registry-server" containerID="cri-o://0c3ba073df55d68d444ae1a53438f841ae80ba8f57ef35005a2764708a336392" gracePeriod=30 Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.449567 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876"] Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.457556 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/238b1bf541f8a95c9abe79f21bceddb01ae5db289e998257aff3a432b14q876"] Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.624997 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.648355 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.689026 4989 generic.go:334] "Generic (PLEG): container finished" podID="69501fce-9a39-435d-a689-6b1eddcf7188" containerID="64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852" exitCode=0 Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.689102 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" event={"ID":"69501fce-9a39-435d-a689-6b1eddcf7188","Type":"ContainerDied","Data":"64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852"} Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.689129 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" event={"ID":"69501fce-9a39-435d-a689-6b1eddcf7188","Type":"ContainerDied","Data":"1a2bff07f73140ade3f6b985348f3c9dbed6cf3f6f4a143db52397a2bdc662af"} Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.689153 4989 scope.go:117] "RemoveContainer" containerID="64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.689261 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-58668b4874-gcpk9" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.695485 4989 generic.go:334] "Generic (PLEG): container finished" podID="7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" containerID="f21c6bbd57cdde1d6f5de8d8c8e6fb35fc8adc5b7649f059ad04df3aa24245a5" exitCode=1 Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.695539 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" event={"ID":"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216","Type":"ContainerDied","Data":"f21c6bbd57cdde1d6f5de8d8c8e6fb35fc8adc5b7649f059ad04df3aa24245a5"} Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.696027 4989 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" secret="" err="secret \"galera-openstack-dockercfg-gvwv4\" not found" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.696073 4989 scope.go:117] "RemoveContainer" containerID="f21c6bbd57cdde1d6f5de8d8c8e6fb35fc8adc5b7649f059ad04df3aa24245a5" Dec 13 17:36:35 crc kubenswrapper[4989]: E1213 17:36:35.696470 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystone7987-account-delete-qq7zw_cinder-kuttl-tests(7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216)\"" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" podUID="7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.701320 4989 generic.go:334] "Generic (PLEG): container finished" podID="93cc60c5-709e-474e-81c5-04242bbd0dec" containerID="aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482" exitCode=0 Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.701409 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.701471 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" event={"ID":"93cc60c5-709e-474e-81c5-04242bbd0dec","Type":"ContainerDied","Data":"aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482"} Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.701503 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv" event={"ID":"93cc60c5-709e-474e-81c5-04242bbd0dec","Type":"ContainerDied","Data":"8467999a1210df50fc886ab6d8a484b810067fab4cbb5b75ed86cfbfebf65fc3"} Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.705494 4989 generic.go:334] "Generic (PLEG): container finished" podID="f814e529-71ee-4f12-8057-ebfe2cd78b97" containerID="0c3ba073df55d68d444ae1a53438f841ae80ba8f57ef35005a2764708a336392" exitCode=0 Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.705542 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-zjpg6" event={"ID":"f814e529-71ee-4f12-8057-ebfe2cd78b97","Type":"ContainerDied","Data":"0c3ba073df55d68d444ae1a53438f841ae80ba8f57ef35005a2764708a336392"} Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.720247 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d9pr\" (UniqueName: \"kubernetes.io/projected/69501fce-9a39-435d-a689-6b1eddcf7188-kube-api-access-4d9pr\") pod \"69501fce-9a39-435d-a689-6b1eddcf7188\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.720334 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-fernet-keys\") pod \"69501fce-9a39-435d-a689-6b1eddcf7188\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.720387 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-apiservice-cert\") pod \"93cc60c5-709e-474e-81c5-04242bbd0dec\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.720417 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-webhook-cert\") pod \"93cc60c5-709e-474e-81c5-04242bbd0dec\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.720544 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bwfh\" (UniqueName: \"kubernetes.io/projected/93cc60c5-709e-474e-81c5-04242bbd0dec-kube-api-access-6bwfh\") pod \"93cc60c5-709e-474e-81c5-04242bbd0dec\" (UID: \"93cc60c5-709e-474e-81c5-04242bbd0dec\") " Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.720600 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-credential-keys\") pod \"69501fce-9a39-435d-a689-6b1eddcf7188\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.720629 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-scripts\") pod \"69501fce-9a39-435d-a689-6b1eddcf7188\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.720664 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-config-data\") pod \"69501fce-9a39-435d-a689-6b1eddcf7188\" (UID: \"69501fce-9a39-435d-a689-6b1eddcf7188\") " Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.727444 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69501fce-9a39-435d-a689-6b1eddcf7188-kube-api-access-4d9pr" (OuterVolumeSpecName: "kube-api-access-4d9pr") pod "69501fce-9a39-435d-a689-6b1eddcf7188" (UID: "69501fce-9a39-435d-a689-6b1eddcf7188"). InnerVolumeSpecName "kube-api-access-4d9pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.728440 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "69501fce-9a39-435d-a689-6b1eddcf7188" (UID: "69501fce-9a39-435d-a689-6b1eddcf7188"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.729044 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93cc60c5-709e-474e-81c5-04242bbd0dec-kube-api-access-6bwfh" (OuterVolumeSpecName: "kube-api-access-6bwfh") pod "93cc60c5-709e-474e-81c5-04242bbd0dec" (UID: "93cc60c5-709e-474e-81c5-04242bbd0dec"). InnerVolumeSpecName "kube-api-access-6bwfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.730298 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "69501fce-9a39-435d-a689-6b1eddcf7188" (UID: "69501fce-9a39-435d-a689-6b1eddcf7188"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.730395 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "93cc60c5-709e-474e-81c5-04242bbd0dec" (UID: "93cc60c5-709e-474e-81c5-04242bbd0dec"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.730849 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-scripts" (OuterVolumeSpecName: "scripts") pod "69501fce-9a39-435d-a689-6b1eddcf7188" (UID: "69501fce-9a39-435d-a689-6b1eddcf7188"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.731294 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "93cc60c5-709e-474e-81c5-04242bbd0dec" (UID: "93cc60c5-709e-474e-81c5-04242bbd0dec"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.736101 4989 scope.go:117] "RemoveContainer" containerID="64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852" Dec 13 17:36:35 crc kubenswrapper[4989]: E1213 17:36:35.739211 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852\": container with ID starting with 64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852 not found: ID does not exist" containerID="64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.739261 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852"} err="failed to get container status \"64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852\": rpc error: code = NotFound desc = could not find container \"64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852\": container with ID starting with 64ad491238e3d386971bc9ca9133382dd78c5ef107ced404a2952d1dc67ac852 not found: ID does not exist" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.739318 4989 scope.go:117] "RemoveContainer" containerID="c676121bddd5b9c87e39c6c8a793e13114bf825bde93b16b72a4b92d2738456a" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.751505 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-config-data" (OuterVolumeSpecName: "config-data") pod "69501fce-9a39-435d-a689-6b1eddcf7188" (UID: "69501fce-9a39-435d-a689-6b1eddcf7188"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.775986 4989 scope.go:117] "RemoveContainer" containerID="aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.785056 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-zjpg6" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.797471 4989 scope.go:117] "RemoveContainer" containerID="aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482" Dec 13 17:36:35 crc kubenswrapper[4989]: E1213 17:36:35.806654 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482\": container with ID starting with aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482 not found: ID does not exist" containerID="aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.806695 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482"} err="failed to get container status \"aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482\": rpc error: code = NotFound desc = could not find container \"aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482\": container with ID starting with aefb2a24bb591fcd6cbeac5dce92aa7fb2225dafafebc1f4901e3af104e57482 not found: ID does not exist" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.821819 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.821876 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-config-data\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.821890 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d9pr\" (UniqueName: \"kubernetes.io/projected/69501fce-9a39-435d-a689-6b1eddcf7188-kube-api-access-4d9pr\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.821904 4989 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.821915 4989 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.821926 4989 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/93cc60c5-709e-474e-81c5-04242bbd0dec-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.821939 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bwfh\" (UniqueName: \"kubernetes.io/projected/93cc60c5-709e-474e-81c5-04242bbd0dec-kube-api-access-6bwfh\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.821951 4989 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/69501fce-9a39-435d-a689-6b1eddcf7188-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.923312 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vwkg\" (UniqueName: \"kubernetes.io/projected/f814e529-71ee-4f12-8057-ebfe2cd78b97-kube-api-access-6vwkg\") pod \"f814e529-71ee-4f12-8057-ebfe2cd78b97\" (UID: \"f814e529-71ee-4f12-8057-ebfe2cd78b97\") " Dec 13 17:36:35 crc kubenswrapper[4989]: I1213 17:36:35.926428 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f814e529-71ee-4f12-8057-ebfe2cd78b97-kube-api-access-6vwkg" (OuterVolumeSpecName: "kube-api-access-6vwkg") pod "f814e529-71ee-4f12-8057-ebfe2cd78b97" (UID: "f814e529-71ee-4f12-8057-ebfe2cd78b97"). InnerVolumeSpecName "kube-api-access-6vwkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.024283 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vwkg\" (UniqueName: \"kubernetes.io/projected/f814e529-71ee-4f12-8057-ebfe2cd78b97-kube-api-access-6vwkg\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.026252 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb43b605-721a-4dae-98c8-eeabd32ba579" path="/var/lib/kubelet/pods/bb43b605-721a-4dae-98c8-eeabd32ba579/volumes" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.027108 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec03663b-83e2-452f-8229-daa7887134be" path="/var/lib/kubelet/pods/ec03663b-83e2-452f-8229-daa7887134be/volumes" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.028270 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-58668b4874-gcpk9"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.028305 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-58668b4874-gcpk9"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.038049 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.042522 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-56ddbfcbfc-bclmv"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.292207 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.436224 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-plugins\") pod \"78da11d7-96b7-4cfd-b190-f34622427b21\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.436288 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78da11d7-96b7-4cfd-b190-f34622427b21-plugins-conf\") pod \"78da11d7-96b7-4cfd-b190-f34622427b21\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.436313 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-confd\") pod \"78da11d7-96b7-4cfd-b190-f34622427b21\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.436346 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78da11d7-96b7-4cfd-b190-f34622427b21-pod-info\") pod \"78da11d7-96b7-4cfd-b190-f34622427b21\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.436397 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78da11d7-96b7-4cfd-b190-f34622427b21-erlang-cookie-secret\") pod \"78da11d7-96b7-4cfd-b190-f34622427b21\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.436447 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-erlang-cookie\") pod \"78da11d7-96b7-4cfd-b190-f34622427b21\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.436509 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzctk\" (UniqueName: \"kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-kube-api-access-rzctk\") pod \"78da11d7-96b7-4cfd-b190-f34622427b21\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.436656 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\") pod \"78da11d7-96b7-4cfd-b190-f34622427b21\" (UID: \"78da11d7-96b7-4cfd-b190-f34622427b21\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.437535 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "78da11d7-96b7-4cfd-b190-f34622427b21" (UID: "78da11d7-96b7-4cfd-b190-f34622427b21"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.437923 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78da11d7-96b7-4cfd-b190-f34622427b21-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "78da11d7-96b7-4cfd-b190-f34622427b21" (UID: "78da11d7-96b7-4cfd-b190-f34622427b21"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.442626 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "78da11d7-96b7-4cfd-b190-f34622427b21" (UID: "78da11d7-96b7-4cfd-b190-f34622427b21"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.446223 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/78da11d7-96b7-4cfd-b190-f34622427b21-pod-info" (OuterVolumeSpecName: "pod-info") pod "78da11d7-96b7-4cfd-b190-f34622427b21" (UID: "78da11d7-96b7-4cfd-b190-f34622427b21"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.448128 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78da11d7-96b7-4cfd-b190-f34622427b21-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "78da11d7-96b7-4cfd-b190-f34622427b21" (UID: "78da11d7-96b7-4cfd-b190-f34622427b21"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.453465 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29" (OuterVolumeSpecName: "persistence") pod "78da11d7-96b7-4cfd-b190-f34622427b21" (UID: "78da11d7-96b7-4cfd-b190-f34622427b21"). InnerVolumeSpecName "pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.453809 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-kube-api-access-rzctk" (OuterVolumeSpecName: "kube-api-access-rzctk") pod "78da11d7-96b7-4cfd-b190-f34622427b21" (UID: "78da11d7-96b7-4cfd-b190-f34622427b21"). InnerVolumeSpecName "kube-api-access-rzctk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.501687 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "78da11d7-96b7-4cfd-b190-f34622427b21" (UID: "78da11d7-96b7-4cfd-b190-f34622427b21"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.538221 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzctk\" (UniqueName: \"kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-kube-api-access-rzctk\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.538268 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\") on node \"crc\" " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.538282 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.538294 4989 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78da11d7-96b7-4cfd-b190-f34622427b21-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.538303 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.538313 4989 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78da11d7-96b7-4cfd-b190-f34622427b21-pod-info\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.538323 4989 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78da11d7-96b7-4cfd-b190-f34622427b21-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.538334 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78da11d7-96b7-4cfd-b190-f34622427b21-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.539617 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.552406 4989 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.552710 4989 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29") on node "crc" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.614421 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-0" podUID="91eeba9c-4026-408d-9547-3bb58ae605dd" containerName="galera" containerID="cri-o://b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4" gracePeriod=26 Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.639532 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-operator-scripts\") pod \"b50061aa-30a3-4145-bca8-94c802594dc4\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.639581 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-kolla-config\") pod \"b50061aa-30a3-4145-bca8-94c802594dc4\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.639604 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-generated\") pod \"b50061aa-30a3-4145-bca8-94c802594dc4\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.639632 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7bww\" (UniqueName: \"kubernetes.io/projected/b50061aa-30a3-4145-bca8-94c802594dc4-kube-api-access-l7bww\") pod \"b50061aa-30a3-4145-bca8-94c802594dc4\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.639858 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-default\") pod \"b50061aa-30a3-4145-bca8-94c802594dc4\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.639879 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"b50061aa-30a3-4145-bca8-94c802594dc4\" (UID: \"b50061aa-30a3-4145-bca8-94c802594dc4\") " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.640133 4989 reconciler_common.go:293] "Volume detached for volume \"pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fe59cabb-4034-4e2e-9f2c-60de0e6f0e29\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.640315 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "b50061aa-30a3-4145-bca8-94c802594dc4" (UID: "b50061aa-30a3-4145-bca8-94c802594dc4"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.640552 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "b50061aa-30a3-4145-bca8-94c802594dc4" (UID: "b50061aa-30a3-4145-bca8-94c802594dc4"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.640659 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "b50061aa-30a3-4145-bca8-94c802594dc4" (UID: "b50061aa-30a3-4145-bca8-94c802594dc4"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.641388 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b50061aa-30a3-4145-bca8-94c802594dc4" (UID: "b50061aa-30a3-4145-bca8-94c802594dc4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.643363 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b50061aa-30a3-4145-bca8-94c802594dc4-kube-api-access-l7bww" (OuterVolumeSpecName: "kube-api-access-l7bww") pod "b50061aa-30a3-4145-bca8-94c802594dc4" (UID: "b50061aa-30a3-4145-bca8-94c802594dc4"). InnerVolumeSpecName "kube-api-access-l7bww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.648605 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "mysql-db") pod "b50061aa-30a3-4145-bca8-94c802594dc4" (UID: "b50061aa-30a3-4145-bca8-94c802594dc4"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.712396 4989 generic.go:334] "Generic (PLEG): container finished" podID="78da11d7-96b7-4cfd-b190-f34622427b21" containerID="9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb" exitCode=0 Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.712453 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"78da11d7-96b7-4cfd-b190-f34622427b21","Type":"ContainerDied","Data":"9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb"} Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.712479 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"78da11d7-96b7-4cfd-b190-f34622427b21","Type":"ContainerDied","Data":"73b5371443f3cb2d3401456c60cbb5b79e6a6c8cd3216369158aee6396f71e35"} Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.712498 4989 scope.go:117] "RemoveContainer" containerID="9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.712628 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.715606 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-zjpg6" event={"ID":"f814e529-71ee-4f12-8057-ebfe2cd78b97","Type":"ContainerDied","Data":"b8cca2e7bc17f710d977a79df1bb0ea068c0593096627fc1dba347c8a3524867"} Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.715911 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-zjpg6" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.717526 4989 generic.go:334] "Generic (PLEG): container finished" podID="b50061aa-30a3-4145-bca8-94c802594dc4" containerID="d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c" exitCode=0 Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.717574 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.717620 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"b50061aa-30a3-4145-bca8-94c802594dc4","Type":"ContainerDied","Data":"d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c"} Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.717654 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"b50061aa-30a3-4145-bca8-94c802594dc4","Type":"ContainerDied","Data":"d89dc977938d99ccded63ad6416b99bb492b32eda5f4ea10635f6d3f862cd994"} Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.721218 4989 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" secret="" err="secret \"galera-openstack-dockercfg-gvwv4\" not found" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.721266 4989 scope.go:117] "RemoveContainer" containerID="f21c6bbd57cdde1d6f5de8d8c8e6fb35fc8adc5b7649f059ad04df3aa24245a5" Dec 13 17:36:36 crc kubenswrapper[4989]: E1213 17:36:36.721508 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystone7987-account-delete-qq7zw_cinder-kuttl-tests(7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216)\"" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" podUID="7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.730190 4989 scope.go:117] "RemoveContainer" containerID="c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.741236 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.741410 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.741484 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.741536 4989 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b50061aa-30a3-4145-bca8-94c802594dc4-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.741585 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b50061aa-30a3-4145-bca8-94c802594dc4-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.741649 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7bww\" (UniqueName: \"kubernetes.io/projected/b50061aa-30a3-4145-bca8-94c802594dc4-kube-api-access-l7bww\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.752459 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-index-zjpg6"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.753077 4989 scope.go:117] "RemoveContainer" containerID="9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb" Dec 13 17:36:36 crc kubenswrapper[4989]: E1213 17:36:36.753473 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb\": container with ID starting with 9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb not found: ID does not exist" containerID="9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.753504 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb"} err="failed to get container status \"9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb\": rpc error: code = NotFound desc = could not find container \"9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb\": container with ID starting with 9516e55a43b87c8a1f9e14a4dda3fc67fcb084211ce877dc30f85c69b971cffb not found: ID does not exist" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.753526 4989 scope.go:117] "RemoveContainer" containerID="c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622" Dec 13 17:36:36 crc kubenswrapper[4989]: E1213 17:36:36.753904 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622\": container with ID starting with c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622 not found: ID does not exist" containerID="c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.753928 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622"} err="failed to get container status \"c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622\": rpc error: code = NotFound desc = could not find container \"c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622\": container with ID starting with c796d84ad176cde57c45be601e1e7ca9e6d9d960aa7e8b05f80b3f5320b83622 not found: ID does not exist" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.753943 4989 scope.go:117] "RemoveContainer" containerID="0c3ba073df55d68d444ae1a53438f841ae80ba8f57ef35005a2764708a336392" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.754237 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.764606 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/cinder-operator-index-zjpg6"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.784775 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.784972 4989 scope.go:117] "RemoveContainer" containerID="d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.791613 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.798798 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.810239 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.815321 4989 scope.go:117] "RemoveContainer" containerID="34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.822726 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-88nfm"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.830315 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-88nfm"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.843263 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.844587 4989 scope.go:117] "RemoveContainer" containerID="d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c" Dec 13 17:36:36 crc kubenswrapper[4989]: E1213 17:36:36.845855 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c\": container with ID starting with d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c not found: ID does not exist" containerID="d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.845896 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c"} err="failed to get container status \"d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c\": rpc error: code = NotFound desc = could not find container \"d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c\": container with ID starting with d153af62740d3a4338fe0e1af5b17aadf8df48652edc8f41c5fda8a8dd3f893c not found: ID does not exist" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.845923 4989 scope.go:117] "RemoveContainer" containerID="34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec" Dec 13 17:36:36 crc kubenswrapper[4989]: E1213 17:36:36.847025 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec\": container with ID starting with 34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec not found: ID does not exist" containerID="34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.847058 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec"} err="failed to get container status \"34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec\": rpc error: code = NotFound desc = could not find container \"34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec\": container with ID starting with 34c4446b295b1611f68a0f1180df49ddeacf5f7b6b467712e425b7c4da15f7ec not found: ID does not exist" Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.853277 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone7987-account-delete-qq7zw"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.857978 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-7987-account-create-update-jtc97"] Dec 13 17:36:36 crc kubenswrapper[4989]: I1213 17:36:36.862405 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-7987-account-create-update-jtc97"] Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.278079 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.350683 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-kolla-config\") pod \"91eeba9c-4026-408d-9547-3bb58ae605dd\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.350779 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"91eeba9c-4026-408d-9547-3bb58ae605dd\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.350845 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-generated\") pod \"91eeba9c-4026-408d-9547-3bb58ae605dd\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.350892 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7chb\" (UniqueName: \"kubernetes.io/projected/91eeba9c-4026-408d-9547-3bb58ae605dd-kube-api-access-q7chb\") pod \"91eeba9c-4026-408d-9547-3bb58ae605dd\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.350936 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-default\") pod \"91eeba9c-4026-408d-9547-3bb58ae605dd\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.351010 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-operator-scripts\") pod \"91eeba9c-4026-408d-9547-3bb58ae605dd\" (UID: \"91eeba9c-4026-408d-9547-3bb58ae605dd\") " Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.351086 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "91eeba9c-4026-408d-9547-3bb58ae605dd" (UID: "91eeba9c-4026-408d-9547-3bb58ae605dd"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.351285 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "91eeba9c-4026-408d-9547-3bb58ae605dd" (UID: "91eeba9c-4026-408d-9547-3bb58ae605dd"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.351306 4989 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:37 crc kubenswrapper[4989]: E1213 17:36:37.351356 4989 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Dec 13 17:36:37 crc kubenswrapper[4989]: E1213 17:36:37.351394 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts podName:7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216 nodeName:}" failed. No retries permitted until 2025-12-13 17:36:41.351382057 +0000 UTC m=+1215.957829195 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts") pod "keystone7987-account-delete-qq7zw" (UID: "7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216") : configmap "openstack-scripts" not found Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.351988 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "91eeba9c-4026-408d-9547-3bb58ae605dd" (UID: "91eeba9c-4026-408d-9547-3bb58ae605dd"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.352840 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "91eeba9c-4026-408d-9547-3bb58ae605dd" (UID: "91eeba9c-4026-408d-9547-3bb58ae605dd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.356007 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91eeba9c-4026-408d-9547-3bb58ae605dd-kube-api-access-q7chb" (OuterVolumeSpecName: "kube-api-access-q7chb") pod "91eeba9c-4026-408d-9547-3bb58ae605dd" (UID: "91eeba9c-4026-408d-9547-3bb58ae605dd"). InnerVolumeSpecName "kube-api-access-q7chb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.359389 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "91eeba9c-4026-408d-9547-3bb58ae605dd" (UID: "91eeba9c-4026-408d-9547-3bb58ae605dd"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.452302 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.452334 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.452344 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7chb\" (UniqueName: \"kubernetes.io/projected/91eeba9c-4026-408d-9547-3bb58ae605dd-kube-api-access-q7chb\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.452353 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.452362 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91eeba9c-4026-408d-9547-3bb58ae605dd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.462930 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.553683 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.738170 4989 generic.go:334] "Generic (PLEG): container finished" podID="91eeba9c-4026-408d-9547-3bb58ae605dd" containerID="b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4" exitCode=0 Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.738227 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.738235 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"91eeba9c-4026-408d-9547-3bb58ae605dd","Type":"ContainerDied","Data":"b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4"} Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.738316 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"91eeba9c-4026-408d-9547-3bb58ae605dd","Type":"ContainerDied","Data":"34d5cc57105ffa0cae7f18c0be8a48edd2c7bb7f927e152406a6f5cf911fd741"} Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.738344 4989 scope.go:117] "RemoveContainer" containerID="b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.775969 4989 scope.go:117] "RemoveContainer" containerID="9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.807180 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.817015 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.828150 4989 scope.go:117] "RemoveContainer" containerID="b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4" Dec 13 17:36:37 crc kubenswrapper[4989]: E1213 17:36:37.828597 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4\": container with ID starting with b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4 not found: ID does not exist" containerID="b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.828653 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4"} err="failed to get container status \"b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4\": rpc error: code = NotFound desc = could not find container \"b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4\": container with ID starting with b18761f8ff7b6dc7a70b2b244f98b796af687fcd7d193bd23b63219341b504c4 not found: ID does not exist" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.828684 4989 scope.go:117] "RemoveContainer" containerID="9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517" Dec 13 17:36:37 crc kubenswrapper[4989]: E1213 17:36:37.829050 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517\": container with ID starting with 9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517 not found: ID does not exist" containerID="9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517" Dec 13 17:36:37 crc kubenswrapper[4989]: I1213 17:36:37.829096 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517"} err="failed to get container status \"9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517\": rpc error: code = NotFound desc = could not find container \"9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517\": container with ID starting with 9f320115ca58c58381195a4b6a37853bc95b5f8902b890953b741f40f0434517 not found: ID does not exist" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.022587 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="692c8262-bf86-44d0-aa9d-6c03509d3763" path="/var/lib/kubelet/pods/692c8262-bf86-44d0-aa9d-6c03509d3763/volumes" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.022910 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.023264 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69501fce-9a39-435d-a689-6b1eddcf7188" path="/var/lib/kubelet/pods/69501fce-9a39-435d-a689-6b1eddcf7188/volumes" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.024075 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78da11d7-96b7-4cfd-b190-f34622427b21" path="/var/lib/kubelet/pods/78da11d7-96b7-4cfd-b190-f34622427b21/volumes" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.025426 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91eeba9c-4026-408d-9547-3bb58ae605dd" path="/var/lib/kubelet/pods/91eeba9c-4026-408d-9547-3bb58ae605dd/volumes" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.026096 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93cc60c5-709e-474e-81c5-04242bbd0dec" path="/var/lib/kubelet/pods/93cc60c5-709e-474e-81c5-04242bbd0dec/volumes" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.026925 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b50061aa-30a3-4145-bca8-94c802594dc4" path="/var/lib/kubelet/pods/b50061aa-30a3-4145-bca8-94c802594dc4/volumes" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.028617 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd652c3b-9226-437c-a04e-961498888534" path="/var/lib/kubelet/pods/dd652c3b-9226-437c-a04e-961498888534/volumes" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.029306 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f814e529-71ee-4f12-8057-ebfe2cd78b97" path="/var/lib/kubelet/pods/f814e529-71ee-4f12-8057-ebfe2cd78b97/volumes" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.175912 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b8tp\" (UniqueName: \"kubernetes.io/projected/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-kube-api-access-2b8tp\") pod \"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216\" (UID: \"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216\") " Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.176028 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts\") pod \"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216\" (UID: \"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216\") " Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.177380 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" (UID: "7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.181205 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-kube-api-access-2b8tp" (OuterVolumeSpecName: "kube-api-access-2b8tp") pod "7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" (UID: "7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216"). InnerVolumeSpecName "kube-api-access-2b8tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.278267 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.278323 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b8tp\" (UniqueName: \"kubernetes.io/projected/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216-kube-api-access-2b8tp\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.429542 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6757956774-t4phl"] Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.430066 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" podUID="61c47c47-cdac-40dc-8644-e2cdbccc560e" containerName="manager" containerID="cri-o://f643bfc403e39a79d0f6a354f4d18832db63b89a48d82571f80f70ea650083c4" gracePeriod=10 Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.707782 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-9x9cg"] Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.707975 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-9x9cg" podUID="c8d4b0de-4254-46f6-99dd-2e077312b011" containerName="registry-server" containerID="cri-o://1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1" gracePeriod=30 Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.734221 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx"] Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.751783 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/835551ba8f27f4fd61e1b05ebed5cb285496b645cbb6fd0ac403227c85b47hx"] Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.757218 4989 generic.go:334] "Generic (PLEG): container finished" podID="61c47c47-cdac-40dc-8644-e2cdbccc560e" containerID="f643bfc403e39a79d0f6a354f4d18832db63b89a48d82571f80f70ea650083c4" exitCode=0 Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.757276 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" event={"ID":"61c47c47-cdac-40dc-8644-e2cdbccc560e","Type":"ContainerDied","Data":"f643bfc403e39a79d0f6a354f4d18832db63b89a48d82571f80f70ea650083c4"} Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.770512 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" event={"ID":"7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216","Type":"ContainerDied","Data":"120531dbd988e3ec454a4b8fbf402f2c247c7edaa2f3271e7526a055622e3e15"} Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.770596 4989 scope.go:117] "RemoveContainer" containerID="f21c6bbd57cdde1d6f5de8d8c8e6fb35fc8adc5b7649f059ad04df3aa24245a5" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.770759 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone7987-account-delete-qq7zw" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.815904 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone7987-account-delete-qq7zw"] Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.820479 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone7987-account-delete-qq7zw"] Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.829208 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.829278 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.859920 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.986643 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn2cq\" (UniqueName: \"kubernetes.io/projected/61c47c47-cdac-40dc-8644-e2cdbccc560e-kube-api-access-xn2cq\") pod \"61c47c47-cdac-40dc-8644-e2cdbccc560e\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.986730 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-webhook-cert\") pod \"61c47c47-cdac-40dc-8644-e2cdbccc560e\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.986883 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-apiservice-cert\") pod \"61c47c47-cdac-40dc-8644-e2cdbccc560e\" (UID: \"61c47c47-cdac-40dc-8644-e2cdbccc560e\") " Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.992210 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "61c47c47-cdac-40dc-8644-e2cdbccc560e" (UID: "61c47c47-cdac-40dc-8644-e2cdbccc560e"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.992963 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "61c47c47-cdac-40dc-8644-e2cdbccc560e" (UID: "61c47c47-cdac-40dc-8644-e2cdbccc560e"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:38 crc kubenswrapper[4989]: I1213 17:36:38.994093 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61c47c47-cdac-40dc-8644-e2cdbccc560e-kube-api-access-xn2cq" (OuterVolumeSpecName: "kube-api-access-xn2cq") pod "61c47c47-cdac-40dc-8644-e2cdbccc560e" (UID: "61c47c47-cdac-40dc-8644-e2cdbccc560e"). InnerVolumeSpecName "kube-api-access-xn2cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.088754 4989 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.088803 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn2cq\" (UniqueName: \"kubernetes.io/projected/61c47c47-cdac-40dc-8644-e2cdbccc560e-kube-api-access-xn2cq\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.088814 4989 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/61c47c47-cdac-40dc-8644-e2cdbccc560e-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.120777 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-9x9cg" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.189713 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpbvp\" (UniqueName: \"kubernetes.io/projected/c8d4b0de-4254-46f6-99dd-2e077312b011-kube-api-access-fpbvp\") pod \"c8d4b0de-4254-46f6-99dd-2e077312b011\" (UID: \"c8d4b0de-4254-46f6-99dd-2e077312b011\") " Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.193180 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8d4b0de-4254-46f6-99dd-2e077312b011-kube-api-access-fpbvp" (OuterVolumeSpecName: "kube-api-access-fpbvp") pod "c8d4b0de-4254-46f6-99dd-2e077312b011" (UID: "c8d4b0de-4254-46f6-99dd-2e077312b011"). InnerVolumeSpecName "kube-api-access-fpbvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.291035 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpbvp\" (UniqueName: \"kubernetes.io/projected/c8d4b0de-4254-46f6-99dd-2e077312b011-kube-api-access-fpbvp\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.777776 4989 generic.go:334] "Generic (PLEG): container finished" podID="c8d4b0de-4254-46f6-99dd-2e077312b011" containerID="1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1" exitCode=0 Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.777863 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-9x9cg" event={"ID":"c8d4b0de-4254-46f6-99dd-2e077312b011","Type":"ContainerDied","Data":"1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1"} Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.777894 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-9x9cg" event={"ID":"c8d4b0de-4254-46f6-99dd-2e077312b011","Type":"ContainerDied","Data":"080aad60729212c42dc5aa3472ba223493e9aa0fd612954addccd7ca13432aca"} Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.777910 4989 scope.go:117] "RemoveContainer" containerID="1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.777992 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-9x9cg" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.784247 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" event={"ID":"61c47c47-cdac-40dc-8644-e2cdbccc560e","Type":"ContainerDied","Data":"654c23dbcb19cc26f432e022889861dad23d134cf151b3ab6c1196120184f979"} Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.784323 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-6757956774-t4phl" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.809083 4989 scope.go:117] "RemoveContainer" containerID="1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1" Dec 13 17:36:39 crc kubenswrapper[4989]: E1213 17:36:39.809595 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1\": container with ID starting with 1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1 not found: ID does not exist" containerID="1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.809641 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1"} err="failed to get container status \"1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1\": rpc error: code = NotFound desc = could not find container \"1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1\": container with ID starting with 1059515f13f65819d7cf766e8a4c178bab218e330dd740a94e83a8cd84dcb8e1 not found: ID does not exist" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.809673 4989 scope.go:117] "RemoveContainer" containerID="f643bfc403e39a79d0f6a354f4d18832db63b89a48d82571f80f70ea650083c4" Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.809946 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-9x9cg"] Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.814669 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-9x9cg"] Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.823018 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6757956774-t4phl"] Dec 13 17:36:39 crc kubenswrapper[4989]: I1213 17:36:39.829193 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6757956774-t4phl"] Dec 13 17:36:40 crc kubenswrapper[4989]: I1213 17:36:40.021954 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4146fa85-1ad6-492e-b6a8-71db9e308953" path="/var/lib/kubelet/pods/4146fa85-1ad6-492e-b6a8-71db9e308953/volumes" Dec 13 17:36:40 crc kubenswrapper[4989]: I1213 17:36:40.022995 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61c47c47-cdac-40dc-8644-e2cdbccc560e" path="/var/lib/kubelet/pods/61c47c47-cdac-40dc-8644-e2cdbccc560e/volumes" Dec 13 17:36:40 crc kubenswrapper[4989]: I1213 17:36:40.023686 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" path="/var/lib/kubelet/pods/7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216/volumes" Dec 13 17:36:40 crc kubenswrapper[4989]: I1213 17:36:40.025309 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8d4b0de-4254-46f6-99dd-2e077312b011" path="/var/lib/kubelet/pods/c8d4b0de-4254-46f6-99dd-2e077312b011/volumes" Dec 13 17:36:40 crc kubenswrapper[4989]: I1213 17:36:40.628997 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg"] Dec 13 17:36:40 crc kubenswrapper[4989]: I1213 17:36:40.629246 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" podUID="7334147d-9cc6-47cc-b164-56a971e3e82b" containerName="operator" containerID="cri-o://b448f0bd85df0ca5d9a877c2a09c90416ca3aca7d5fa9bb8c8e23c6be64bfb93" gracePeriod=10 Dec 13 17:36:40 crc kubenswrapper[4989]: I1213 17:36:40.822474 4989 generic.go:334] "Generic (PLEG): container finished" podID="7334147d-9cc6-47cc-b164-56a971e3e82b" containerID="b448f0bd85df0ca5d9a877c2a09c90416ca3aca7d5fa9bb8c8e23c6be64bfb93" exitCode=0 Dec 13 17:36:40 crc kubenswrapper[4989]: I1213 17:36:40.822557 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" event={"ID":"7334147d-9cc6-47cc-b164-56a971e3e82b","Type":"ContainerDied","Data":"b448f0bd85df0ca5d9a877c2a09c90416ca3aca7d5fa9bb8c8e23c6be64bfb93"} Dec 13 17:36:40 crc kubenswrapper[4989]: I1213 17:36:40.958274 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-59pgz"] Dec 13 17:36:40 crc kubenswrapper[4989]: I1213 17:36:40.958687 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" podUID="f0c708a3-be93-496d-85d5-fdedb653aa0d" containerName="registry-server" containerID="cri-o://cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff" gracePeriod=30 Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.011372 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm"] Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.011425 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590sdmqm"] Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.135503 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.217689 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjvbs\" (UniqueName: \"kubernetes.io/projected/7334147d-9cc6-47cc-b164-56a971e3e82b-kube-api-access-jjvbs\") pod \"7334147d-9cc6-47cc-b164-56a971e3e82b\" (UID: \"7334147d-9cc6-47cc-b164-56a971e3e82b\") " Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.223851 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7334147d-9cc6-47cc-b164-56a971e3e82b-kube-api-access-jjvbs" (OuterVolumeSpecName: "kube-api-access-jjvbs") pod "7334147d-9cc6-47cc-b164-56a971e3e82b" (UID: "7334147d-9cc6-47cc-b164-56a971e3e82b"). InnerVolumeSpecName "kube-api-access-jjvbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.283238 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.319521 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjvbs\" (UniqueName: \"kubernetes.io/projected/7334147d-9cc6-47cc-b164-56a971e3e82b-kube-api-access-jjvbs\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.420451 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knfzn\" (UniqueName: \"kubernetes.io/projected/f0c708a3-be93-496d-85d5-fdedb653aa0d-kube-api-access-knfzn\") pod \"f0c708a3-be93-496d-85d5-fdedb653aa0d\" (UID: \"f0c708a3-be93-496d-85d5-fdedb653aa0d\") " Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.423410 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0c708a3-be93-496d-85d5-fdedb653aa0d-kube-api-access-knfzn" (OuterVolumeSpecName: "kube-api-access-knfzn") pod "f0c708a3-be93-496d-85d5-fdedb653aa0d" (UID: "f0c708a3-be93-496d-85d5-fdedb653aa0d"). InnerVolumeSpecName "kube-api-access-knfzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.521802 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knfzn\" (UniqueName: \"kubernetes.io/projected/f0c708a3-be93-496d-85d5-fdedb653aa0d-kube-api-access-knfzn\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.831850 4989 generic.go:334] "Generic (PLEG): container finished" podID="f0c708a3-be93-496d-85d5-fdedb653aa0d" containerID="cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff" exitCode=0 Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.831902 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" event={"ID":"f0c708a3-be93-496d-85d5-fdedb653aa0d","Type":"ContainerDied","Data":"cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff"} Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.831940 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" event={"ID":"f0c708a3-be93-496d-85d5-fdedb653aa0d","Type":"ContainerDied","Data":"24617ac4319df9250811f04c6b64b0ef0a5132ae7477f785ebc8326f64c15b8b"} Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.831943 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-59pgz" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.831962 4989 scope.go:117] "RemoveContainer" containerID="cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.834314 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" event={"ID":"7334147d-9cc6-47cc-b164-56a971e3e82b","Type":"ContainerDied","Data":"96777a515a4f3d1198bbc0ba3306b85398abb5e4768f1761918d90f097e22088"} Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.834436 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.856276 4989 scope.go:117] "RemoveContainer" containerID="cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff" Dec 13 17:36:41 crc kubenswrapper[4989]: E1213 17:36:41.856820 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff\": container with ID starting with cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff not found: ID does not exist" containerID="cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.856889 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff"} err="failed to get container status \"cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff\": rpc error: code = NotFound desc = could not find container \"cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff\": container with ID starting with cb9652e37066d5ed55d358beb0f859f31963ed4f98ee57537442faf7024a35ff not found: ID does not exist" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.856915 4989 scope.go:117] "RemoveContainer" containerID="b448f0bd85df0ca5d9a877c2a09c90416ca3aca7d5fa9bb8c8e23c6be64bfb93" Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.868197 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-59pgz"] Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.875555 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-59pgz"] Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.881801 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg"] Dec 13 17:36:41 crc kubenswrapper[4989]: I1213 17:36:41.886486 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-k55sg"] Dec 13 17:36:42 crc kubenswrapper[4989]: I1213 17:36:42.024580 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ac1af50-c321-49bf-9270-21c2a4afea6a" path="/var/lib/kubelet/pods/5ac1af50-c321-49bf-9270-21c2a4afea6a/volumes" Dec 13 17:36:42 crc kubenswrapper[4989]: I1213 17:36:42.025314 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7334147d-9cc6-47cc-b164-56a971e3e82b" path="/var/lib/kubelet/pods/7334147d-9cc6-47cc-b164-56a971e3e82b/volumes" Dec 13 17:36:42 crc kubenswrapper[4989]: I1213 17:36:42.025877 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0c708a3-be93-496d-85d5-fdedb653aa0d" path="/var/lib/kubelet/pods/f0c708a3-be93-496d-85d5-fdedb653aa0d/volumes" Dec 13 17:36:43 crc kubenswrapper[4989]: I1213 17:36:43.722302 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6"] Dec 13 17:36:43 crc kubenswrapper[4989]: I1213 17:36:43.722537 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" podUID="fbe7e3ff-0c94-4738-9dfc-577938b357d2" containerName="manager" containerID="cri-o://505d75199ebd00bcb7d5e147020cb3c303472d6032b5f7e84d0071a795e7a877" gracePeriod=10 Dec 13 17:36:43 crc kubenswrapper[4989]: I1213 17:36:43.866573 4989 generic.go:334] "Generic (PLEG): container finished" podID="fbe7e3ff-0c94-4738-9dfc-577938b357d2" containerID="505d75199ebd00bcb7d5e147020cb3c303472d6032b5f7e84d0071a795e7a877" exitCode=0 Dec 13 17:36:43 crc kubenswrapper[4989]: I1213 17:36:43.866947 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" event={"ID":"fbe7e3ff-0c94-4738-9dfc-577938b357d2","Type":"ContainerDied","Data":"505d75199ebd00bcb7d5e147020cb3c303472d6032b5f7e84d0071a795e7a877"} Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.030148 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-48dg5"] Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.030358 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-48dg5" podUID="b62f0d30-924c-4e82-9204-1e32c6f4008b" containerName="registry-server" containerID="cri-o://8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227" gracePeriod=30 Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.054645 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q"] Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.058590 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/a60eba3f147c507a234cb91c07236b96c6864b9bfe0f3df79307df7f14z5t4q"] Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.216022 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.359748 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrcdb\" (UniqueName: \"kubernetes.io/projected/fbe7e3ff-0c94-4738-9dfc-577938b357d2-kube-api-access-jrcdb\") pod \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.359884 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-webhook-cert\") pod \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.359920 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-apiservice-cert\") pod \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\" (UID: \"fbe7e3ff-0c94-4738-9dfc-577938b357d2\") " Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.366341 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "fbe7e3ff-0c94-4738-9dfc-577938b357d2" (UID: "fbe7e3ff-0c94-4738-9dfc-577938b357d2"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.379866 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbe7e3ff-0c94-4738-9dfc-577938b357d2-kube-api-access-jrcdb" (OuterVolumeSpecName: "kube-api-access-jrcdb") pod "fbe7e3ff-0c94-4738-9dfc-577938b357d2" (UID: "fbe7e3ff-0c94-4738-9dfc-577938b357d2"). InnerVolumeSpecName "kube-api-access-jrcdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.382630 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "fbe7e3ff-0c94-4738-9dfc-577938b357d2" (UID: "fbe7e3ff-0c94-4738-9dfc-577938b357d2"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.402166 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-48dg5" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.460903 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfdz6\" (UniqueName: \"kubernetes.io/projected/b62f0d30-924c-4e82-9204-1e32c6f4008b-kube-api-access-qfdz6\") pod \"b62f0d30-924c-4e82-9204-1e32c6f4008b\" (UID: \"b62f0d30-924c-4e82-9204-1e32c6f4008b\") " Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.461206 4989 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.461220 4989 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fbe7e3ff-0c94-4738-9dfc-577938b357d2-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.461231 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrcdb\" (UniqueName: \"kubernetes.io/projected/fbe7e3ff-0c94-4738-9dfc-577938b357d2-kube-api-access-jrcdb\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.464069 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b62f0d30-924c-4e82-9204-1e32c6f4008b-kube-api-access-qfdz6" (OuterVolumeSpecName: "kube-api-access-qfdz6") pod "b62f0d30-924c-4e82-9204-1e32c6f4008b" (UID: "b62f0d30-924c-4e82-9204-1e32c6f4008b"). InnerVolumeSpecName "kube-api-access-qfdz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.563015 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfdz6\" (UniqueName: \"kubernetes.io/projected/b62f0d30-924c-4e82-9204-1e32c6f4008b-kube-api-access-qfdz6\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.875396 4989 generic.go:334] "Generic (PLEG): container finished" podID="b62f0d30-924c-4e82-9204-1e32c6f4008b" containerID="8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227" exitCode=0 Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.875500 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-48dg5" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.875511 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-48dg5" event={"ID":"b62f0d30-924c-4e82-9204-1e32c6f4008b","Type":"ContainerDied","Data":"8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227"} Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.875615 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-48dg5" event={"ID":"b62f0d30-924c-4e82-9204-1e32c6f4008b","Type":"ContainerDied","Data":"3fdce0456abc1bd6334ebe71e77ae54a76820557f4132a7bb15804e56852b1a9"} Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.875646 4989 scope.go:117] "RemoveContainer" containerID="8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.877464 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" event={"ID":"fbe7e3ff-0c94-4738-9dfc-577938b357d2","Type":"ContainerDied","Data":"d97a167ced38ed806aa2cb9b0f4f8369d635b3aaf8d67dd78c13ff1908a74696"} Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.877517 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.896300 4989 scope.go:117] "RemoveContainer" containerID="8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227" Dec 13 17:36:44 crc kubenswrapper[4989]: E1213 17:36:44.896776 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227\": container with ID starting with 8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227 not found: ID does not exist" containerID="8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.896828 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227"} err="failed to get container status \"8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227\": rpc error: code = NotFound desc = could not find container \"8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227\": container with ID starting with 8e176a7321918800426971232fd81536aec03c1a6af3968b8cf3329ba14f3227 not found: ID does not exist" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.896853 4989 scope.go:117] "RemoveContainer" containerID="505d75199ebd00bcb7d5e147020cb3c303472d6032b5f7e84d0071a795e7a877" Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.912852 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-48dg5"] Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.924412 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-48dg5"] Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.940107 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6"] Dec 13 17:36:44 crc kubenswrapper[4989]: I1213 17:36:44.941277 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6db5fd65d9-ct8s6"] Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.305643 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44"] Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.305901 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" podUID="c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f" containerName="manager" containerID="cri-o://4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94" gracePeriod=10 Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.645747 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-zcs27"] Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.645976 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-zcs27" podUID="3a635a9e-69b4-4b82-b9fe-207ac9e5184e" containerName="registry-server" containerID="cri-o://6ff595da2c13ecdb90d6cdd271e0ce54dcc432b791ae68955cedd7c483848ff0" gracePeriod=30 Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.711859 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s"] Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.717896 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/5eeaa1363b8e228040d7ca2b19aec73ff370c3e7c3f2a3fbdddaf0de6e8ck7s"] Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.877398 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.910728 4989 generic.go:334] "Generic (PLEG): container finished" podID="3a635a9e-69b4-4b82-b9fe-207ac9e5184e" containerID="6ff595da2c13ecdb90d6cdd271e0ce54dcc432b791ae68955cedd7c483848ff0" exitCode=0 Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.910783 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zcs27" event={"ID":"3a635a9e-69b4-4b82-b9fe-207ac9e5184e","Type":"ContainerDied","Data":"6ff595da2c13ecdb90d6cdd271e0ce54dcc432b791ae68955cedd7c483848ff0"} Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.912603 4989 generic.go:334] "Generic (PLEG): container finished" podID="c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f" containerID="4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94" exitCode=0 Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.912627 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" event={"ID":"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f","Type":"ContainerDied","Data":"4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94"} Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.912642 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" event={"ID":"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f","Type":"ContainerDied","Data":"3e30ad3d228b29de0c5a37d1d836a13cd6016bccb1d2ca1be8c5c167bc4f8bda"} Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.912658 4989 scope.go:117] "RemoveContainer" containerID="4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94" Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.912741 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44" Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.929647 4989 scope.go:117] "RemoveContainer" containerID="4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94" Dec 13 17:36:45 crc kubenswrapper[4989]: E1213 17:36:45.930131 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94\": container with ID starting with 4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94 not found: ID does not exist" containerID="4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94" Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.930157 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94"} err="failed to get container status \"4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94\": rpc error: code = NotFound desc = could not find container \"4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94\": container with ID starting with 4d28055c6b09c2f2bd150bef727501ece7744eed0261f16d02b7aad29191fe94 not found: ID does not exist" Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.982869 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-apiservice-cert\") pod \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.982921 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-webhook-cert\") pod \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.983046 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj9kb\" (UniqueName: \"kubernetes.io/projected/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-kube-api-access-rj9kb\") pod \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\" (UID: \"c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f\") " Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.986579 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f" (UID: "c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.986818 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f" (UID: "c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:36:45 crc kubenswrapper[4989]: I1213 17:36:45.987012 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-kube-api-access-rj9kb" (OuterVolumeSpecName: "kube-api-access-rj9kb") pod "c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f" (UID: "c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f"). InnerVolumeSpecName "kube-api-access-rj9kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.019397 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zcs27" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.024633 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="492342cb-48b9-4cc7-ad59-1c4a815e665b" path="/var/lib/kubelet/pods/492342cb-48b9-4cc7-ad59-1c4a815e665b/volumes" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.025680 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fa079b8-c658-4ba8-b0a4-0034c70e2d7d" path="/var/lib/kubelet/pods/5fa079b8-c658-4ba8-b0a4-0034c70e2d7d/volumes" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.026515 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b62f0d30-924c-4e82-9204-1e32c6f4008b" path="/var/lib/kubelet/pods/b62f0d30-924c-4e82-9204-1e32c6f4008b/volumes" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.027734 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbe7e3ff-0c94-4738-9dfc-577938b357d2" path="/var/lib/kubelet/pods/fbe7e3ff-0c94-4738-9dfc-577938b357d2/volumes" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.085586 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5d2d\" (UniqueName: \"kubernetes.io/projected/3a635a9e-69b4-4b82-b9fe-207ac9e5184e-kube-api-access-l5d2d\") pod \"3a635a9e-69b4-4b82-b9fe-207ac9e5184e\" (UID: \"3a635a9e-69b4-4b82-b9fe-207ac9e5184e\") " Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.085934 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj9kb\" (UniqueName: \"kubernetes.io/projected/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-kube-api-access-rj9kb\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.085953 4989 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.085962 4989 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.088444 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a635a9e-69b4-4b82-b9fe-207ac9e5184e-kube-api-access-l5d2d" (OuterVolumeSpecName: "kube-api-access-l5d2d") pod "3a635a9e-69b4-4b82-b9fe-207ac9e5184e" (UID: "3a635a9e-69b4-4b82-b9fe-207ac9e5184e"). InnerVolumeSpecName "kube-api-access-l5d2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.187644 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5d2d\" (UniqueName: \"kubernetes.io/projected/3a635a9e-69b4-4b82-b9fe-207ac9e5184e-kube-api-access-l5d2d\") on node \"crc\" DevicePath \"\"" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.228589 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44"] Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.233366 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-57575d5b7f-sxp44"] Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.922052 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-zcs27" event={"ID":"3a635a9e-69b4-4b82-b9fe-207ac9e5184e","Type":"ContainerDied","Data":"db651614d4764b6bc695582240d1a5ac01d3d28c50005f21e7b8fb1a9ec73062"} Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.922105 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-zcs27" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.922123 4989 scope.go:117] "RemoveContainer" containerID="6ff595da2c13ecdb90d6cdd271e0ce54dcc432b791ae68955cedd7c483848ff0" Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.952499 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-zcs27"] Dec 13 17:36:46 crc kubenswrapper[4989]: I1213 17:36:46.956512 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-zcs27"] Dec 13 17:36:48 crc kubenswrapper[4989]: I1213 17:36:48.027289 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a635a9e-69b4-4b82-b9fe-207ac9e5184e" path="/var/lib/kubelet/pods/3a635a9e-69b4-4b82-b9fe-207ac9e5184e/volumes" Dec 13 17:36:48 crc kubenswrapper[4989]: I1213 17:36:48.028696 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f" path="/var/lib/kubelet/pods/c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f/volumes" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.468230 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gspn6/must-gather-88znf"] Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.468916 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0c708a3-be93-496d-85d5-fdedb653aa0d" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.468928 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0c708a3-be93-496d-85d5-fdedb653aa0d" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.468940 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f814e529-71ee-4f12-8057-ebfe2cd78b97" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.468946 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f814e529-71ee-4f12-8057-ebfe2cd78b97" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.468956 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe7e3ff-0c94-4738-9dfc-577938b357d2" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.468964 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe7e3ff-0c94-4738-9dfc-577938b357d2" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.468971 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="probe" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.468976 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d04e42-14da-4e12-ad36-826e3bc3a6d3" containerName="probe" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.468984 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69501fce-9a39-435d-a689-6b1eddcf7188" containerName="keystone-api" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.468989 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="69501fce-9a39-435d-a689-6b1eddcf7188" containerName="keystone-api" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469001 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a635a9e-69b4-4b82-b9fe-207ac9e5184e" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469007 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a635a9e-69b4-4b82-b9fe-207ac9e5184e" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469014 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91eeba9c-4026-408d-9547-3bb58ae605dd" containerName="galera" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469020 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91eeba9c-4026-408d-9547-3bb58ae605dd" containerName="galera" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469028 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" containerName="mariadb-account-delete" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469033 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" containerName="mariadb-account-delete" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469039 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb43b605-721a-4dae-98c8-eeabd32ba579" containerName="memcached" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469045 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb43b605-721a-4dae-98c8-eeabd32ba579" containerName="memcached" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469052 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78da11d7-96b7-4cfd-b190-f34622427b21" containerName="setup-container" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469058 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="78da11d7-96b7-4cfd-b190-f34622427b21" containerName="setup-container" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469064 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62f0d30-924c-4e82-9204-1e32c6f4008b" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469070 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62f0d30-924c-4e82-9204-1e32c6f4008b" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469078 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469083 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469093 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93cc60c5-709e-474e-81c5-04242bbd0dec" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469098 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="93cc60c5-709e-474e-81c5-04242bbd0dec" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469104 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c540cd7f-4b50-4a71-9e51-0687156ee1b2" containerName="galera" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469110 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c540cd7f-4b50-4a71-9e51-0687156ee1b2" containerName="galera" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469119 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78da11d7-96b7-4cfd-b190-f34622427b21" containerName="rabbitmq" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469124 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="78da11d7-96b7-4cfd-b190-f34622427b21" containerName="rabbitmq" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469131 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91eeba9c-4026-408d-9547-3bb58ae605dd" containerName="mysql-bootstrap" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469137 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91eeba9c-4026-408d-9547-3bb58ae605dd" containerName="mysql-bootstrap" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469143 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d4b0de-4254-46f6-99dd-2e077312b011" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469149 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d4b0de-4254-46f6-99dd-2e077312b011" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469155 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b50061aa-30a3-4145-bca8-94c802594dc4" containerName="mysql-bootstrap" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469161 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b50061aa-30a3-4145-bca8-94c802594dc4" containerName="mysql-bootstrap" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469170 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61c47c47-cdac-40dc-8644-e2cdbccc560e" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469176 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="61c47c47-cdac-40dc-8644-e2cdbccc560e" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469185 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b50061aa-30a3-4145-bca8-94c802594dc4" containerName="galera" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469191 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b50061aa-30a3-4145-bca8-94c802594dc4" containerName="galera" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469197 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" containerName="mariadb-account-delete" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469203 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" containerName="mariadb-account-delete" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469210 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c540cd7f-4b50-4a71-9e51-0687156ee1b2" containerName="mysql-bootstrap" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469218 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c540cd7f-4b50-4a71-9e51-0687156ee1b2" containerName="mysql-bootstrap" Dec 13 17:36:59 crc kubenswrapper[4989]: E1213 17:36:59.469225 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7334147d-9cc6-47cc-b164-56a971e3e82b" containerName="operator" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469230 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7334147d-9cc6-47cc-b164-56a971e3e82b" containerName="operator" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469322 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8d4b0de-4254-46f6-99dd-2e077312b011" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469333 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f814e529-71ee-4f12-8057-ebfe2cd78b97" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469339 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b62f0d30-924c-4e82-9204-1e32c6f4008b" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469349 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0c708a3-be93-496d-85d5-fdedb653aa0d" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469357 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c540cd7f-4b50-4a71-9e51-0687156ee1b2" containerName="galera" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469366 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="61c47c47-cdac-40dc-8644-e2cdbccc560e" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469372 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="78da11d7-96b7-4cfd-b190-f34622427b21" containerName="rabbitmq" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469379 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91eeba9c-4026-408d-9547-3bb58ae605dd" containerName="galera" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469388 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb43b605-721a-4dae-98c8-eeabd32ba579" containerName="memcached" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469395 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a635a9e-69b4-4b82-b9fe-207ac9e5184e" containerName="registry-server" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469404 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b50061aa-30a3-4145-bca8-94c802594dc4" containerName="galera" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469412 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" containerName="mariadb-account-delete" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469420 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbe7e3ff-0c94-4738-9dfc-577938b357d2" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469427 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="69501fce-9a39-435d-a689-6b1eddcf7188" containerName="keystone-api" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469433 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="93cc60c5-709e-474e-81c5-04242bbd0dec" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469442 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7334147d-9cc6-47cc-b164-56a971e3e82b" containerName="operator" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469448 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1fdf4ee-0738-4bb6-a4e6-2f460ed5530f" containerName="manager" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.469626 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd57ccc-b4af-47c1-a0cb-bcd4dbbd5216" containerName="mariadb-account-delete" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.470037 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gspn6/must-gather-88znf" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.476757 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gspn6"/"openshift-service-ca.crt" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.478181 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gspn6"/"kube-root-ca.crt" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.485527 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gspn6/must-gather-88znf"] Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.580015 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6195da6d-e28a-4f8e-9ddc-49ef134b660a-must-gather-output\") pod \"must-gather-88znf\" (UID: \"6195da6d-e28a-4f8e-9ddc-49ef134b660a\") " pod="openshift-must-gather-gspn6/must-gather-88znf" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.580147 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p2sj\" (UniqueName: \"kubernetes.io/projected/6195da6d-e28a-4f8e-9ddc-49ef134b660a-kube-api-access-6p2sj\") pod \"must-gather-88znf\" (UID: \"6195da6d-e28a-4f8e-9ddc-49ef134b660a\") " pod="openshift-must-gather-gspn6/must-gather-88znf" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.681084 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6195da6d-e28a-4f8e-9ddc-49ef134b660a-must-gather-output\") pod \"must-gather-88znf\" (UID: \"6195da6d-e28a-4f8e-9ddc-49ef134b660a\") " pod="openshift-must-gather-gspn6/must-gather-88znf" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.681175 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p2sj\" (UniqueName: \"kubernetes.io/projected/6195da6d-e28a-4f8e-9ddc-49ef134b660a-kube-api-access-6p2sj\") pod \"must-gather-88znf\" (UID: \"6195da6d-e28a-4f8e-9ddc-49ef134b660a\") " pod="openshift-must-gather-gspn6/must-gather-88znf" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.681499 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6195da6d-e28a-4f8e-9ddc-49ef134b660a-must-gather-output\") pod \"must-gather-88znf\" (UID: \"6195da6d-e28a-4f8e-9ddc-49ef134b660a\") " pod="openshift-must-gather-gspn6/must-gather-88znf" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.701885 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p2sj\" (UniqueName: \"kubernetes.io/projected/6195da6d-e28a-4f8e-9ddc-49ef134b660a-kube-api-access-6p2sj\") pod \"must-gather-88znf\" (UID: \"6195da6d-e28a-4f8e-9ddc-49ef134b660a\") " pod="openshift-must-gather-gspn6/must-gather-88znf" Dec 13 17:36:59 crc kubenswrapper[4989]: I1213 17:36:59.786233 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gspn6/must-gather-88znf" Dec 13 17:37:00 crc kubenswrapper[4989]: I1213 17:37:00.198627 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gspn6/must-gather-88znf"] Dec 13 17:37:01 crc kubenswrapper[4989]: I1213 17:37:01.036688 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gspn6/must-gather-88znf" event={"ID":"6195da6d-e28a-4f8e-9ddc-49ef134b660a","Type":"ContainerStarted","Data":"3cc4d67981460f629ae371adceadea72d224b8731bc5ab0531c01a7e31f15eb6"} Dec 13 17:37:04 crc kubenswrapper[4989]: I1213 17:37:04.059722 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gspn6/must-gather-88znf" event={"ID":"6195da6d-e28a-4f8e-9ddc-49ef134b660a","Type":"ContainerStarted","Data":"af73581950f6e70fc85f0c98c5c6e3f2b7b272be5f9f870642faadcad6f4eaf6"} Dec 13 17:37:05 crc kubenswrapper[4989]: I1213 17:37:05.080225 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gspn6/must-gather-88znf" event={"ID":"6195da6d-e28a-4f8e-9ddc-49ef134b660a","Type":"ContainerStarted","Data":"b913c45167c30437277a111bf6796841767d0d56b03703269a6405563a2a2f0c"} Dec 13 17:37:05 crc kubenswrapper[4989]: I1213 17:37:05.106563 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gspn6/must-gather-88znf" podStartSLOduration=2.635748106 podStartE2EDuration="6.106545859s" podCreationTimestamp="2025-12-13 17:36:59 +0000 UTC" firstStartedPulling="2025-12-13 17:37:00.207518288 +0000 UTC m=+1234.813965416" lastFinishedPulling="2025-12-13 17:37:03.678316031 +0000 UTC m=+1238.284763169" observedRunningTime="2025-12-13 17:37:05.100980545 +0000 UTC m=+1239.707427683" watchObservedRunningTime="2025-12-13 17:37:05.106545859 +0000 UTC m=+1239.712992997" Dec 13 17:37:08 crc kubenswrapper[4989]: I1213 17:37:08.829609 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:37:08 crc kubenswrapper[4989]: I1213 17:37:08.829954 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:37:26 crc kubenswrapper[4989]: I1213 17:37:26.869457 4989 scope.go:117] "RemoveContainer" containerID="d6f7a9d8d20124ddb87de8aa31da75a829a05909e735e7459bed3a4ba493c25d" Dec 13 17:37:26 crc kubenswrapper[4989]: I1213 17:37:26.911363 4989 scope.go:117] "RemoveContainer" containerID="66a9891c3b5ca710046e7f8b45890b09bf2086d4cb2cb1a6c247dfeb3a55a95d" Dec 13 17:37:26 crc kubenswrapper[4989]: I1213 17:37:26.938362 4989 scope.go:117] "RemoveContainer" containerID="99408b9a55f861ab6c0e10e8c233f9fe5790378eeba2ea3c6f8a001834572670" Dec 13 17:37:26 crc kubenswrapper[4989]: I1213 17:37:26.952249 4989 scope.go:117] "RemoveContainer" containerID="f5d55af33e4e0fa93a533a9ae15179cf09fb9b6818ef13896df5a8e0ee51230a" Dec 13 17:37:26 crc kubenswrapper[4989]: I1213 17:37:26.973229 4989 scope.go:117] "RemoveContainer" containerID="b2e20b8dc38bf625f7e62af2d3e368be4afc7fbd9cbe17074aa5612c72a3d2f2" Dec 13 17:37:27 crc kubenswrapper[4989]: I1213 17:37:27.002654 4989 scope.go:117] "RemoveContainer" containerID="840ecc0ff6c76d7c717a8a2c4f4c54600b99f2864110036f4a3e12e55ed827f3" Dec 13 17:37:27 crc kubenswrapper[4989]: I1213 17:37:27.020602 4989 scope.go:117] "RemoveContainer" containerID="424109aa8329e2aa5518d5a8bc6bb03cc68221af0e5ec6a7aec49992552e9603" Dec 13 17:37:27 crc kubenswrapper[4989]: I1213 17:37:27.037461 4989 scope.go:117] "RemoveContainer" containerID="74a78c7c620e342a5341cdc87224b5e81a96ccf2a68d1f5cd518274d58301291" Dec 13 17:37:27 crc kubenswrapper[4989]: I1213 17:37:27.059104 4989 scope.go:117] "RemoveContainer" containerID="69bfdcd39b6722f9908d6c327460719cc72bb5f5c69dda1fbcd0eefeee40325d" Dec 13 17:37:27 crc kubenswrapper[4989]: I1213 17:37:27.078483 4989 scope.go:117] "RemoveContainer" containerID="558803f067315776cf7ac1b786343f01938f94576665abad2bc8c725fdfee1f9" Dec 13 17:37:38 crc kubenswrapper[4989]: I1213 17:37:38.829994 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:37:38 crc kubenswrapper[4989]: I1213 17:37:38.830381 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:37:38 crc kubenswrapper[4989]: I1213 17:37:38.830417 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:37:38 crc kubenswrapper[4989]: I1213 17:37:38.830836 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7a7aa4339807f600a727da9523920b27ce8d8223ea989c8dc88df4c91d1c4743"} pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 17:37:38 crc kubenswrapper[4989]: I1213 17:37:38.830880 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" containerID="cri-o://7a7aa4339807f600a727da9523920b27ce8d8223ea989c8dc88df4c91d1c4743" gracePeriod=600 Dec 13 17:37:39 crc kubenswrapper[4989]: I1213 17:37:39.321913 4989 generic.go:334] "Generic (PLEG): container finished" podID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerID="7a7aa4339807f600a727da9523920b27ce8d8223ea989c8dc88df4c91d1c4743" exitCode=0 Dec 13 17:37:39 crc kubenswrapper[4989]: I1213 17:37:39.321981 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerDied","Data":"7a7aa4339807f600a727da9523920b27ce8d8223ea989c8dc88df4c91d1c4743"} Dec 13 17:37:39 crc kubenswrapper[4989]: I1213 17:37:39.322470 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerStarted","Data":"0b5b47b60f26e4730dadda3c98b1d3052f3ae310472ca5f800e8c41049d96126"} Dec 13 17:37:39 crc kubenswrapper[4989]: I1213 17:37:39.322518 4989 scope.go:117] "RemoveContainer" containerID="f7a34899e26b30f2d8d01fb7d7814d99ead88a61514d27cbbc8a2e72806061a9" Dec 13 17:37:46 crc kubenswrapper[4989]: I1213 17:37:46.889396 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-fkvxp_a134b88a-8fa0-486b-b66d-b1df520f4a6c/control-plane-machine-set-operator/0.log" Dec 13 17:37:47 crc kubenswrapper[4989]: I1213 17:37:47.069745 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8hwgn_7fe4787f-cdd3-44fe-8405-dd6ebc9b455c/kube-rbac-proxy/0.log" Dec 13 17:37:47 crc kubenswrapper[4989]: I1213 17:37:47.085154 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8hwgn_7fe4787f-cdd3-44fe-8405-dd6ebc9b455c/machine-api-operator/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.226156 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-55pxp_c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29/kube-rbac-proxy/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.329032 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-55pxp_c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29/controller/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.449942 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-frr-files/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.619818 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-frr-files/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.629904 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-metrics/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.633478 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-reloader/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.642941 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-reloader/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.830286 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-reloader/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.842422 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-metrics/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.862340 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-frr-files/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.906421 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-metrics/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.992422 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-frr-files/0.log" Dec 13 17:38:02 crc kubenswrapper[4989]: I1213 17:38:02.998520 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-reloader/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.039446 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-metrics/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.098258 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/controller/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.154891 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/frr-metrics/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.229820 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/kube-rbac-proxy/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.267395 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/kube-rbac-proxy-frr/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.367867 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/reloader/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.503140 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-2shrx_7c4da269-7f3c-4101-8776-3c10b7379d9c/frr-k8s-webhook-server/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.647883 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/frr/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.706147 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-64f9694c74-whn2q_1ba2b41f-9366-455a-ab5c-51a199ac8d52/manager/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.778387 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-84cdcb4857-zvzst_9f540b88-5e5f-4420-8b2c-d49f42087943/webhook-server/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.868986 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r5qp2_931ee02f-2887-4831-a0df-94eccdf81672/kube-rbac-proxy/0.log" Dec 13 17:38:03 crc kubenswrapper[4989]: I1213 17:38:03.988673 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r5qp2_931ee02f-2887-4831-a0df-94eccdf81672/speaker/0.log" Dec 13 17:38:27 crc kubenswrapper[4989]: I1213 17:38:27.322528 4989 scope.go:117] "RemoveContainer" containerID="d79bbdd009a893187a04a7dc41238e9bc4c5cd0893bcef281570e76f4720e0cc" Dec 13 17:38:27 crc kubenswrapper[4989]: I1213 17:38:27.346448 4989 scope.go:117] "RemoveContainer" containerID="3ecc834249e0941b8d5a34b4963ecec7ce2f4115f753e77b0bd0be0a8677231a" Dec 13 17:38:27 crc kubenswrapper[4989]: I1213 17:38:27.375713 4989 scope.go:117] "RemoveContainer" containerID="83c90348a987db1c060ab96646093a76fd802d23c8b03282aad61fe4e10bac2a" Dec 13 17:38:28 crc kubenswrapper[4989]: I1213 17:38:28.030518 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/util/0.log" Dec 13 17:38:28 crc kubenswrapper[4989]: I1213 17:38:28.215159 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/util/0.log" Dec 13 17:38:28 crc kubenswrapper[4989]: I1213 17:38:28.234446 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/pull/0.log" Dec 13 17:38:28 crc kubenswrapper[4989]: I1213 17:38:28.250218 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/pull/0.log" Dec 13 17:38:28 crc kubenswrapper[4989]: I1213 17:38:28.468682 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/util/0.log" Dec 13 17:38:28 crc kubenswrapper[4989]: I1213 17:38:28.470393 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/pull/0.log" Dec 13 17:38:28 crc kubenswrapper[4989]: I1213 17:38:28.471147 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/extract/0.log" Dec 13 17:38:28 crc kubenswrapper[4989]: I1213 17:38:28.634646 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-utilities/0.log" Dec 13 17:38:28 crc kubenswrapper[4989]: I1213 17:38:28.834819 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-utilities/0.log" Dec 13 17:38:28 crc kubenswrapper[4989]: I1213 17:38:28.834863 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-content/0.log" Dec 13 17:38:28 crc kubenswrapper[4989]: I1213 17:38:28.900414 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-content/0.log" Dec 13 17:38:29 crc kubenswrapper[4989]: I1213 17:38:29.120300 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-utilities/0.log" Dec 13 17:38:29 crc kubenswrapper[4989]: I1213 17:38:29.155379 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-content/0.log" Dec 13 17:38:29 crc kubenswrapper[4989]: I1213 17:38:29.388116 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-utilities/0.log" Dec 13 17:38:29 crc kubenswrapper[4989]: I1213 17:38:29.434087 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/registry-server/0.log" Dec 13 17:38:29 crc kubenswrapper[4989]: I1213 17:38:29.718040 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-utilities/0.log" Dec 13 17:38:29 crc kubenswrapper[4989]: I1213 17:38:29.747616 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-content/0.log" Dec 13 17:38:29 crc kubenswrapper[4989]: I1213 17:38:29.843050 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-content/0.log" Dec 13 17:38:29 crc kubenswrapper[4989]: I1213 17:38:29.975147 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-utilities/0.log" Dec 13 17:38:30 crc kubenswrapper[4989]: I1213 17:38:30.020751 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-content/0.log" Dec 13 17:38:30 crc kubenswrapper[4989]: I1213 17:38:30.212449 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wl78k_a56dab99-6f88-4da4-af60-913e0b4c3be6/marketplace-operator/0.log" Dec 13 17:38:30 crc kubenswrapper[4989]: I1213 17:38:30.253496 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/registry-server/0.log" Dec 13 17:38:30 crc kubenswrapper[4989]: I1213 17:38:30.328733 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-utilities/0.log" Dec 13 17:38:30 crc kubenswrapper[4989]: I1213 17:38:30.506189 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-content/0.log" Dec 13 17:38:30 crc kubenswrapper[4989]: I1213 17:38:30.506282 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-utilities/0.log" Dec 13 17:38:30 crc kubenswrapper[4989]: I1213 17:38:30.512507 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-content/0.log" Dec 13 17:38:30 crc kubenswrapper[4989]: I1213 17:38:30.765128 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-utilities/0.log" Dec 13 17:38:30 crc kubenswrapper[4989]: I1213 17:38:30.788808 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-content/0.log" Dec 13 17:38:30 crc kubenswrapper[4989]: I1213 17:38:30.829467 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/registry-server/0.log" Dec 13 17:38:30 crc kubenswrapper[4989]: I1213 17:38:30.955159 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-utilities/0.log" Dec 13 17:38:31 crc kubenswrapper[4989]: I1213 17:38:31.145692 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-content/0.log" Dec 13 17:38:31 crc kubenswrapper[4989]: I1213 17:38:31.182399 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-utilities/0.log" Dec 13 17:38:31 crc kubenswrapper[4989]: I1213 17:38:31.189234 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-content/0.log" Dec 13 17:38:31 crc kubenswrapper[4989]: I1213 17:38:31.365573 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-utilities/0.log" Dec 13 17:38:31 crc kubenswrapper[4989]: I1213 17:38:31.378265 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-content/0.log" Dec 13 17:38:31 crc kubenswrapper[4989]: I1213 17:38:31.662581 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/registry-server/0.log" Dec 13 17:39:27 crc kubenswrapper[4989]: I1213 17:39:27.422008 4989 scope.go:117] "RemoveContainer" containerID="0ac6efa75569331505db0a8eaf9130e6b79234a6680e073a55690ca42ff089fd" Dec 13 17:39:27 crc kubenswrapper[4989]: I1213 17:39:27.476732 4989 scope.go:117] "RemoveContainer" containerID="5d311671a6cc64730cce91cf40f7dbedf82c270d3fe28bdbd32d6780f50522ed" Dec 13 17:39:27 crc kubenswrapper[4989]: I1213 17:39:27.501010 4989 scope.go:117] "RemoveContainer" containerID="b204627536e2bd230ccdbce339c40382083b723f0251ec3a5c252ec4509e1e65" Dec 13 17:39:27 crc kubenswrapper[4989]: I1213 17:39:27.536846 4989 scope.go:117] "RemoveContainer" containerID="3bc9eac8dd55bc4b680ff962671d8574d0fa09d888fe7108a89428dab407a2bf" Dec 13 17:39:27 crc kubenswrapper[4989]: I1213 17:39:27.596623 4989 scope.go:117] "RemoveContainer" containerID="0be360a14bddf56a685592fbdda141d9a4ccbefb17d572ca98b8362dc89d57df" Dec 13 17:39:27 crc kubenswrapper[4989]: I1213 17:39:27.624677 4989 scope.go:117] "RemoveContainer" containerID="0112ac7b8b36b6378ffce6ae04f4262d5d3787408857526c0bf270dad7889142" Dec 13 17:39:27 crc kubenswrapper[4989]: I1213 17:39:27.651520 4989 scope.go:117] "RemoveContainer" containerID="4eded5d33a154bfc6fed1cf9d904d9b07585b44d62d7d663caade4661dfdba92" Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.653648 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cnm2p"] Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.655286 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.671941 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cnm2p"] Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.687575 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-utilities\") pod \"certified-operators-cnm2p\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.687626 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-catalog-content\") pod \"certified-operators-cnm2p\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.687672 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t2ht\" (UniqueName: \"kubernetes.io/projected/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-kube-api-access-9t2ht\") pod \"certified-operators-cnm2p\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.788476 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-catalog-content\") pod \"certified-operators-cnm2p\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.788609 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t2ht\" (UniqueName: \"kubernetes.io/projected/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-kube-api-access-9t2ht\") pod \"certified-operators-cnm2p\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.789032 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-utilities\") pod \"certified-operators-cnm2p\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.789058 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-catalog-content\") pod \"certified-operators-cnm2p\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.789411 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-utilities\") pod \"certified-operators-cnm2p\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.813840 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t2ht\" (UniqueName: \"kubernetes.io/projected/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-kube-api-access-9t2ht\") pod \"certified-operators-cnm2p\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:38 crc kubenswrapper[4989]: I1213 17:39:38.986435 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:39 crc kubenswrapper[4989]: I1213 17:39:39.464162 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cnm2p"] Dec 13 17:39:40 crc kubenswrapper[4989]: I1213 17:39:40.062127 4989 generic.go:334] "Generic (PLEG): container finished" podID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" containerID="5612c278c7fd8c136fba0485ac5a4bf3cb1bc2f6111d40df93d92c54906249e5" exitCode=0 Dec 13 17:39:40 crc kubenswrapper[4989]: I1213 17:39:40.062259 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnm2p" event={"ID":"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889","Type":"ContainerDied","Data":"5612c278c7fd8c136fba0485ac5a4bf3cb1bc2f6111d40df93d92c54906249e5"} Dec 13 17:39:40 crc kubenswrapper[4989]: I1213 17:39:40.062572 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnm2p" event={"ID":"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889","Type":"ContainerStarted","Data":"d40ac2ef44648f822e173958929e4405af9b3dafd1e0c1db86c6c8338e4d2a32"} Dec 13 17:39:40 crc kubenswrapper[4989]: I1213 17:39:40.064078 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 17:39:41 crc kubenswrapper[4989]: I1213 17:39:41.070997 4989 generic.go:334] "Generic (PLEG): container finished" podID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" containerID="5521583784d24ed5ee651534f2d0ef0efbb2af4a305079b7164cbb7557781e71" exitCode=0 Dec 13 17:39:41 crc kubenswrapper[4989]: I1213 17:39:41.071055 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnm2p" event={"ID":"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889","Type":"ContainerDied","Data":"5521583784d24ed5ee651534f2d0ef0efbb2af4a305079b7164cbb7557781e71"} Dec 13 17:39:42 crc kubenswrapper[4989]: I1213 17:39:42.081631 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnm2p" event={"ID":"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889","Type":"ContainerStarted","Data":"14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e"} Dec 13 17:39:42 crc kubenswrapper[4989]: I1213 17:39:42.098772 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cnm2p" podStartSLOduration=2.506796334 podStartE2EDuration="4.098753859s" podCreationTimestamp="2025-12-13 17:39:38 +0000 UTC" firstStartedPulling="2025-12-13 17:39:40.063869709 +0000 UTC m=+1394.670316847" lastFinishedPulling="2025-12-13 17:39:41.655827244 +0000 UTC m=+1396.262274372" observedRunningTime="2025-12-13 17:39:42.097406087 +0000 UTC m=+1396.703853235" watchObservedRunningTime="2025-12-13 17:39:42.098753859 +0000 UTC m=+1396.705201017" Dec 13 17:39:44 crc kubenswrapper[4989]: I1213 17:39:44.104528 4989 generic.go:334] "Generic (PLEG): container finished" podID="6195da6d-e28a-4f8e-9ddc-49ef134b660a" containerID="af73581950f6e70fc85f0c98c5c6e3f2b7b272be5f9f870642faadcad6f4eaf6" exitCode=0 Dec 13 17:39:44 crc kubenswrapper[4989]: I1213 17:39:44.105536 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gspn6/must-gather-88znf" event={"ID":"6195da6d-e28a-4f8e-9ddc-49ef134b660a","Type":"ContainerDied","Data":"af73581950f6e70fc85f0c98c5c6e3f2b7b272be5f9f870642faadcad6f4eaf6"} Dec 13 17:39:44 crc kubenswrapper[4989]: I1213 17:39:44.106448 4989 scope.go:117] "RemoveContainer" containerID="af73581950f6e70fc85f0c98c5c6e3f2b7b272be5f9f870642faadcad6f4eaf6" Dec 13 17:39:44 crc kubenswrapper[4989]: I1213 17:39:44.377477 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gspn6_must-gather-88znf_6195da6d-e28a-4f8e-9ddc-49ef134b660a/gather/0.log" Dec 13 17:39:48 crc kubenswrapper[4989]: I1213 17:39:48.987673 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:48 crc kubenswrapper[4989]: I1213 17:39:48.991868 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:49 crc kubenswrapper[4989]: I1213 17:39:49.035979 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:49 crc kubenswrapper[4989]: I1213 17:39:49.194646 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:49 crc kubenswrapper[4989]: I1213 17:39:49.279151 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cnm2p"] Dec 13 17:39:50 crc kubenswrapper[4989]: I1213 17:39:50.978606 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gspn6/must-gather-88znf"] Dec 13 17:39:50 crc kubenswrapper[4989]: I1213 17:39:50.979109 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gspn6/must-gather-88znf" podUID="6195da6d-e28a-4f8e-9ddc-49ef134b660a" containerName="copy" containerID="cri-o://b913c45167c30437277a111bf6796841767d0d56b03703269a6405563a2a2f0c" gracePeriod=2 Dec 13 17:39:50 crc kubenswrapper[4989]: I1213 17:39:50.982527 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gspn6/must-gather-88znf"] Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.153694 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gspn6_must-gather-88znf_6195da6d-e28a-4f8e-9ddc-49ef134b660a/copy/0.log" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.154571 4989 generic.go:334] "Generic (PLEG): container finished" podID="6195da6d-e28a-4f8e-9ddc-49ef134b660a" containerID="b913c45167c30437277a111bf6796841767d0d56b03703269a6405563a2a2f0c" exitCode=143 Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.154812 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cnm2p" podUID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" containerName="registry-server" containerID="cri-o://14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e" gracePeriod=2 Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.359382 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gspn6_must-gather-88znf_6195da6d-e28a-4f8e-9ddc-49ef134b660a/copy/0.log" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.359890 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gspn6/must-gather-88znf" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.471130 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p2sj\" (UniqueName: \"kubernetes.io/projected/6195da6d-e28a-4f8e-9ddc-49ef134b660a-kube-api-access-6p2sj\") pod \"6195da6d-e28a-4f8e-9ddc-49ef134b660a\" (UID: \"6195da6d-e28a-4f8e-9ddc-49ef134b660a\") " Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.471197 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6195da6d-e28a-4f8e-9ddc-49ef134b660a-must-gather-output\") pod \"6195da6d-e28a-4f8e-9ddc-49ef134b660a\" (UID: \"6195da6d-e28a-4f8e-9ddc-49ef134b660a\") " Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.478911 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6195da6d-e28a-4f8e-9ddc-49ef134b660a-kube-api-access-6p2sj" (OuterVolumeSpecName: "kube-api-access-6p2sj") pod "6195da6d-e28a-4f8e-9ddc-49ef134b660a" (UID: "6195da6d-e28a-4f8e-9ddc-49ef134b660a"). InnerVolumeSpecName "kube-api-access-6p2sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.510439 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.551685 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6195da6d-e28a-4f8e-9ddc-49ef134b660a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6195da6d-e28a-4f8e-9ddc-49ef134b660a" (UID: "6195da6d-e28a-4f8e-9ddc-49ef134b660a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.572105 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-utilities\") pod \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.572203 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t2ht\" (UniqueName: \"kubernetes.io/projected/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-kube-api-access-9t2ht\") pod \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.572220 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-catalog-content\") pod \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\" (UID: \"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889\") " Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.572890 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p2sj\" (UniqueName: \"kubernetes.io/projected/6195da6d-e28a-4f8e-9ddc-49ef134b660a-kube-api-access-6p2sj\") on node \"crc\" DevicePath \"\"" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.572908 4989 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6195da6d-e28a-4f8e-9ddc-49ef134b660a-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.573273 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-utilities" (OuterVolumeSpecName: "utilities") pod "dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" (UID: "dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.576907 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-kube-api-access-9t2ht" (OuterVolumeSpecName: "kube-api-access-9t2ht") pod "dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" (UID: "dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889"). InnerVolumeSpecName "kube-api-access-9t2ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.623781 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" (UID: "dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.673821 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.673847 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t2ht\" (UniqueName: \"kubernetes.io/projected/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-kube-api-access-9t2ht\") on node \"crc\" DevicePath \"\"" Dec 13 17:39:51 crc kubenswrapper[4989]: I1213 17:39:51.673858 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.030552 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6195da6d-e28a-4f8e-9ddc-49ef134b660a" path="/var/lib/kubelet/pods/6195da6d-e28a-4f8e-9ddc-49ef134b660a/volumes" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.162598 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gspn6_must-gather-88znf_6195da6d-e28a-4f8e-9ddc-49ef134b660a/copy/0.log" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.163119 4989 scope.go:117] "RemoveContainer" containerID="b913c45167c30437277a111bf6796841767d0d56b03703269a6405563a2a2f0c" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.163281 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gspn6/must-gather-88znf" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.167675 4989 generic.go:334] "Generic (PLEG): container finished" podID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" containerID="14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e" exitCode=0 Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.167725 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnm2p" event={"ID":"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889","Type":"ContainerDied","Data":"14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e"} Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.167756 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cnm2p" event={"ID":"dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889","Type":"ContainerDied","Data":"d40ac2ef44648f822e173958929e4405af9b3dafd1e0c1db86c6c8338e4d2a32"} Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.167937 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cnm2p" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.188930 4989 scope.go:117] "RemoveContainer" containerID="af73581950f6e70fc85f0c98c5c6e3f2b7b272be5f9f870642faadcad6f4eaf6" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.214672 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cnm2p"] Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.224207 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cnm2p"] Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.252655 4989 scope.go:117] "RemoveContainer" containerID="14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.274501 4989 scope.go:117] "RemoveContainer" containerID="5521583784d24ed5ee651534f2d0ef0efbb2af4a305079b7164cbb7557781e71" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.304281 4989 scope.go:117] "RemoveContainer" containerID="5612c278c7fd8c136fba0485ac5a4bf3cb1bc2f6111d40df93d92c54906249e5" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.330309 4989 scope.go:117] "RemoveContainer" containerID="14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e" Dec 13 17:39:52 crc kubenswrapper[4989]: E1213 17:39:52.331570 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e\": container with ID starting with 14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e not found: ID does not exist" containerID="14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.331646 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e"} err="failed to get container status \"14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e\": rpc error: code = NotFound desc = could not find container \"14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e\": container with ID starting with 14a1dca28b7569e8f722ec69c1429a339030bdd17cc551965c6c95e6b7f9ab0e not found: ID does not exist" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.331703 4989 scope.go:117] "RemoveContainer" containerID="5521583784d24ed5ee651534f2d0ef0efbb2af4a305079b7164cbb7557781e71" Dec 13 17:39:52 crc kubenswrapper[4989]: E1213 17:39:52.332157 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5521583784d24ed5ee651534f2d0ef0efbb2af4a305079b7164cbb7557781e71\": container with ID starting with 5521583784d24ed5ee651534f2d0ef0efbb2af4a305079b7164cbb7557781e71 not found: ID does not exist" containerID="5521583784d24ed5ee651534f2d0ef0efbb2af4a305079b7164cbb7557781e71" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.332197 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5521583784d24ed5ee651534f2d0ef0efbb2af4a305079b7164cbb7557781e71"} err="failed to get container status \"5521583784d24ed5ee651534f2d0ef0efbb2af4a305079b7164cbb7557781e71\": rpc error: code = NotFound desc = could not find container \"5521583784d24ed5ee651534f2d0ef0efbb2af4a305079b7164cbb7557781e71\": container with ID starting with 5521583784d24ed5ee651534f2d0ef0efbb2af4a305079b7164cbb7557781e71 not found: ID does not exist" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.332230 4989 scope.go:117] "RemoveContainer" containerID="5612c278c7fd8c136fba0485ac5a4bf3cb1bc2f6111d40df93d92c54906249e5" Dec 13 17:39:52 crc kubenswrapper[4989]: E1213 17:39:52.332593 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5612c278c7fd8c136fba0485ac5a4bf3cb1bc2f6111d40df93d92c54906249e5\": container with ID starting with 5612c278c7fd8c136fba0485ac5a4bf3cb1bc2f6111d40df93d92c54906249e5 not found: ID does not exist" containerID="5612c278c7fd8c136fba0485ac5a4bf3cb1bc2f6111d40df93d92c54906249e5" Dec 13 17:39:52 crc kubenswrapper[4989]: I1213 17:39:52.332653 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5612c278c7fd8c136fba0485ac5a4bf3cb1bc2f6111d40df93d92c54906249e5"} err="failed to get container status \"5612c278c7fd8c136fba0485ac5a4bf3cb1bc2f6111d40df93d92c54906249e5\": rpc error: code = NotFound desc = could not find container \"5612c278c7fd8c136fba0485ac5a4bf3cb1bc2f6111d40df93d92c54906249e5\": container with ID starting with 5612c278c7fd8c136fba0485ac5a4bf3cb1bc2f6111d40df93d92c54906249e5 not found: ID does not exist" Dec 13 17:39:54 crc kubenswrapper[4989]: I1213 17:39:54.025633 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" path="/var/lib/kubelet/pods/dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889/volumes" Dec 13 17:40:08 crc kubenswrapper[4989]: I1213 17:40:08.830418 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:40:08 crc kubenswrapper[4989]: I1213 17:40:08.831597 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:40:27 crc kubenswrapper[4989]: I1213 17:40:27.724980 4989 scope.go:117] "RemoveContainer" containerID="f706b390351e2d289287322d78fbdbcda39572588c81ccc338aee0aa910a6ca9" Dec 13 17:40:27 crc kubenswrapper[4989]: I1213 17:40:27.741253 4989 scope.go:117] "RemoveContainer" containerID="1e5dbc5fdc6f743c02662f552524d0621038235624f8c4c568c58474afdf9afd" Dec 13 17:40:27 crc kubenswrapper[4989]: I1213 17:40:27.757876 4989 scope.go:117] "RemoveContainer" containerID="47e1793bb65ee7bf294065057b894ba482b4c88971a1740483f994fc1f0f6172" Dec 13 17:40:38 crc kubenswrapper[4989]: I1213 17:40:38.829299 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:40:38 crc kubenswrapper[4989]: I1213 17:40:38.829895 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:41:03 crc kubenswrapper[4989]: I1213 17:41:03.925429 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mgn8w"] Dec 13 17:41:03 crc kubenswrapper[4989]: E1213 17:41:03.926192 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" containerName="extract-content" Dec 13 17:41:03 crc kubenswrapper[4989]: I1213 17:41:03.926211 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" containerName="extract-content" Dec 13 17:41:03 crc kubenswrapper[4989]: E1213 17:41:03.926226 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6195da6d-e28a-4f8e-9ddc-49ef134b660a" containerName="gather" Dec 13 17:41:03 crc kubenswrapper[4989]: I1213 17:41:03.926324 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6195da6d-e28a-4f8e-9ddc-49ef134b660a" containerName="gather" Dec 13 17:41:03 crc kubenswrapper[4989]: E1213 17:41:03.926346 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" containerName="registry-server" Dec 13 17:41:03 crc kubenswrapper[4989]: I1213 17:41:03.926358 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" containerName="registry-server" Dec 13 17:41:03 crc kubenswrapper[4989]: E1213 17:41:03.926371 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" containerName="extract-utilities" Dec 13 17:41:03 crc kubenswrapper[4989]: I1213 17:41:03.926382 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" containerName="extract-utilities" Dec 13 17:41:03 crc kubenswrapper[4989]: E1213 17:41:03.926410 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6195da6d-e28a-4f8e-9ddc-49ef134b660a" containerName="copy" Dec 13 17:41:03 crc kubenswrapper[4989]: I1213 17:41:03.926420 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6195da6d-e28a-4f8e-9ddc-49ef134b660a" containerName="copy" Dec 13 17:41:03 crc kubenswrapper[4989]: I1213 17:41:03.926547 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd1ff4c6-c9ca-41d9-9058-f3e81ccbb889" containerName="registry-server" Dec 13 17:41:03 crc kubenswrapper[4989]: I1213 17:41:03.926566 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6195da6d-e28a-4f8e-9ddc-49ef134b660a" containerName="gather" Dec 13 17:41:03 crc kubenswrapper[4989]: I1213 17:41:03.926579 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6195da6d-e28a-4f8e-9ddc-49ef134b660a" containerName="copy" Dec 13 17:41:03 crc kubenswrapper[4989]: I1213 17:41:03.927532 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:03 crc kubenswrapper[4989]: I1213 17:41:03.938098 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgn8w"] Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.058689 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-catalog-content\") pod \"redhat-marketplace-mgn8w\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.058746 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncgm5\" (UniqueName: \"kubernetes.io/projected/03013fe5-bdd4-45a3-abec-1e32d6a017c1-kube-api-access-ncgm5\") pod \"redhat-marketplace-mgn8w\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.058949 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-utilities\") pod \"redhat-marketplace-mgn8w\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.159854 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-catalog-content\") pod \"redhat-marketplace-mgn8w\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.159914 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncgm5\" (UniqueName: \"kubernetes.io/projected/03013fe5-bdd4-45a3-abec-1e32d6a017c1-kube-api-access-ncgm5\") pod \"redhat-marketplace-mgn8w\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.159949 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-utilities\") pod \"redhat-marketplace-mgn8w\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.160356 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-catalog-content\") pod \"redhat-marketplace-mgn8w\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.160424 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-utilities\") pod \"redhat-marketplace-mgn8w\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.178577 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncgm5\" (UniqueName: \"kubernetes.io/projected/03013fe5-bdd4-45a3-abec-1e32d6a017c1-kube-api-access-ncgm5\") pod \"redhat-marketplace-mgn8w\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.260461 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.470731 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgn8w"] Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.676039 4989 generic.go:334] "Generic (PLEG): container finished" podID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" containerID="b301322e82e1a4c2a18d699e32b8eac5d3e309cc0928919cd9e29ca7b4e08b97" exitCode=0 Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.676092 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgn8w" event={"ID":"03013fe5-bdd4-45a3-abec-1e32d6a017c1","Type":"ContainerDied","Data":"b301322e82e1a4c2a18d699e32b8eac5d3e309cc0928919cd9e29ca7b4e08b97"} Dec 13 17:41:04 crc kubenswrapper[4989]: I1213 17:41:04.676124 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgn8w" event={"ID":"03013fe5-bdd4-45a3-abec-1e32d6a017c1","Type":"ContainerStarted","Data":"fdd65600391792eae59d1bc324758cd64326a33b1ee9823be86245b57c658721"} Dec 13 17:41:05 crc kubenswrapper[4989]: I1213 17:41:05.684681 4989 generic.go:334] "Generic (PLEG): container finished" podID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" containerID="d298135a67d450c59bcd1dd8554305d0e7d67e8a2a720755257630fdc45ab14f" exitCode=0 Dec 13 17:41:05 crc kubenswrapper[4989]: I1213 17:41:05.684814 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgn8w" event={"ID":"03013fe5-bdd4-45a3-abec-1e32d6a017c1","Type":"ContainerDied","Data":"d298135a67d450c59bcd1dd8554305d0e7d67e8a2a720755257630fdc45ab14f"} Dec 13 17:41:06 crc kubenswrapper[4989]: I1213 17:41:06.696390 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgn8w" event={"ID":"03013fe5-bdd4-45a3-abec-1e32d6a017c1","Type":"ContainerStarted","Data":"5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1"} Dec 13 17:41:06 crc kubenswrapper[4989]: I1213 17:41:06.723039 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mgn8w" podStartSLOduration=2.179345752 podStartE2EDuration="3.723022476s" podCreationTimestamp="2025-12-13 17:41:03 +0000 UTC" firstStartedPulling="2025-12-13 17:41:04.677901611 +0000 UTC m=+1479.284348749" lastFinishedPulling="2025-12-13 17:41:06.221578335 +0000 UTC m=+1480.828025473" observedRunningTime="2025-12-13 17:41:06.72120547 +0000 UTC m=+1481.327652608" watchObservedRunningTime="2025-12-13 17:41:06.723022476 +0000 UTC m=+1481.329469614" Dec 13 17:41:08 crc kubenswrapper[4989]: I1213 17:41:08.829415 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:41:08 crc kubenswrapper[4989]: I1213 17:41:08.829487 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:41:08 crc kubenswrapper[4989]: I1213 17:41:08.829534 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:41:08 crc kubenswrapper[4989]: I1213 17:41:08.830149 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b5b47b60f26e4730dadda3c98b1d3052f3ae310472ca5f800e8c41049d96126"} pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 17:41:08 crc kubenswrapper[4989]: I1213 17:41:08.830196 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" containerID="cri-o://0b5b47b60f26e4730dadda3c98b1d3052f3ae310472ca5f800e8c41049d96126" gracePeriod=600 Dec 13 17:41:09 crc kubenswrapper[4989]: I1213 17:41:09.722511 4989 generic.go:334] "Generic (PLEG): container finished" podID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerID="0b5b47b60f26e4730dadda3c98b1d3052f3ae310472ca5f800e8c41049d96126" exitCode=0 Dec 13 17:41:09 crc kubenswrapper[4989]: I1213 17:41:09.722563 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerDied","Data":"0b5b47b60f26e4730dadda3c98b1d3052f3ae310472ca5f800e8c41049d96126"} Dec 13 17:41:09 crc kubenswrapper[4989]: I1213 17:41:09.723341 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerStarted","Data":"24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2"} Dec 13 17:41:09 crc kubenswrapper[4989]: I1213 17:41:09.723377 4989 scope.go:117] "RemoveContainer" containerID="7a7aa4339807f600a727da9523920b27ce8d8223ea989c8dc88df4c91d1c4743" Dec 13 17:41:14 crc kubenswrapper[4989]: I1213 17:41:14.260772 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:14 crc kubenswrapper[4989]: I1213 17:41:14.261169 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:14 crc kubenswrapper[4989]: I1213 17:41:14.303657 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:14 crc kubenswrapper[4989]: I1213 17:41:14.804198 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:14 crc kubenswrapper[4989]: I1213 17:41:14.841005 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgn8w"] Dec 13 17:41:16 crc kubenswrapper[4989]: I1213 17:41:16.768985 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mgn8w" podUID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" containerName="registry-server" containerID="cri-o://5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1" gracePeriod=2 Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.085477 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.224938 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncgm5\" (UniqueName: \"kubernetes.io/projected/03013fe5-bdd4-45a3-abec-1e32d6a017c1-kube-api-access-ncgm5\") pod \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.224979 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-catalog-content\") pod \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.225057 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-utilities\") pod \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\" (UID: \"03013fe5-bdd4-45a3-abec-1e32d6a017c1\") " Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.226226 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-utilities" (OuterVolumeSpecName: "utilities") pod "03013fe5-bdd4-45a3-abec-1e32d6a017c1" (UID: "03013fe5-bdd4-45a3-abec-1e32d6a017c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.230299 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03013fe5-bdd4-45a3-abec-1e32d6a017c1-kube-api-access-ncgm5" (OuterVolumeSpecName: "kube-api-access-ncgm5") pod "03013fe5-bdd4-45a3-abec-1e32d6a017c1" (UID: "03013fe5-bdd4-45a3-abec-1e32d6a017c1"). InnerVolumeSpecName "kube-api-access-ncgm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.246467 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03013fe5-bdd4-45a3-abec-1e32d6a017c1" (UID: "03013fe5-bdd4-45a3-abec-1e32d6a017c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.327062 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.327108 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03013fe5-bdd4-45a3-abec-1e32d6a017c1-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.327118 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncgm5\" (UniqueName: \"kubernetes.io/projected/03013fe5-bdd4-45a3-abec-1e32d6a017c1-kube-api-access-ncgm5\") on node \"crc\" DevicePath \"\"" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.777141 4989 generic.go:334] "Generic (PLEG): container finished" podID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" containerID="5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1" exitCode=0 Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.777190 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgn8w" event={"ID":"03013fe5-bdd4-45a3-abec-1e32d6a017c1","Type":"ContainerDied","Data":"5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1"} Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.777209 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mgn8w" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.777238 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mgn8w" event={"ID":"03013fe5-bdd4-45a3-abec-1e32d6a017c1","Type":"ContainerDied","Data":"fdd65600391792eae59d1bc324758cd64326a33b1ee9823be86245b57c658721"} Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.777270 4989 scope.go:117] "RemoveContainer" containerID="5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.798771 4989 scope.go:117] "RemoveContainer" containerID="d298135a67d450c59bcd1dd8554305d0e7d67e8a2a720755257630fdc45ab14f" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.814451 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgn8w"] Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.818706 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mgn8w"] Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.834596 4989 scope.go:117] "RemoveContainer" containerID="b301322e82e1a4c2a18d699e32b8eac5d3e309cc0928919cd9e29ca7b4e08b97" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.855234 4989 scope.go:117] "RemoveContainer" containerID="5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1" Dec 13 17:41:17 crc kubenswrapper[4989]: E1213 17:41:17.856191 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1\": container with ID starting with 5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1 not found: ID does not exist" containerID="5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.856249 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1"} err="failed to get container status \"5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1\": rpc error: code = NotFound desc = could not find container \"5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1\": container with ID starting with 5294512a122ad84ed1540c4941e7a4ac4caa2ab8d571004a766e6d5a4e8361c1 not found: ID does not exist" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.856297 4989 scope.go:117] "RemoveContainer" containerID="d298135a67d450c59bcd1dd8554305d0e7d67e8a2a720755257630fdc45ab14f" Dec 13 17:41:17 crc kubenswrapper[4989]: E1213 17:41:17.858901 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d298135a67d450c59bcd1dd8554305d0e7d67e8a2a720755257630fdc45ab14f\": container with ID starting with d298135a67d450c59bcd1dd8554305d0e7d67e8a2a720755257630fdc45ab14f not found: ID does not exist" containerID="d298135a67d450c59bcd1dd8554305d0e7d67e8a2a720755257630fdc45ab14f" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.858931 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d298135a67d450c59bcd1dd8554305d0e7d67e8a2a720755257630fdc45ab14f"} err="failed to get container status \"d298135a67d450c59bcd1dd8554305d0e7d67e8a2a720755257630fdc45ab14f\": rpc error: code = NotFound desc = could not find container \"d298135a67d450c59bcd1dd8554305d0e7d67e8a2a720755257630fdc45ab14f\": container with ID starting with d298135a67d450c59bcd1dd8554305d0e7d67e8a2a720755257630fdc45ab14f not found: ID does not exist" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.858950 4989 scope.go:117] "RemoveContainer" containerID="b301322e82e1a4c2a18d699e32b8eac5d3e309cc0928919cd9e29ca7b4e08b97" Dec 13 17:41:17 crc kubenswrapper[4989]: E1213 17:41:17.860124 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b301322e82e1a4c2a18d699e32b8eac5d3e309cc0928919cd9e29ca7b4e08b97\": container with ID starting with b301322e82e1a4c2a18d699e32b8eac5d3e309cc0928919cd9e29ca7b4e08b97 not found: ID does not exist" containerID="b301322e82e1a4c2a18d699e32b8eac5d3e309cc0928919cd9e29ca7b4e08b97" Dec 13 17:41:17 crc kubenswrapper[4989]: I1213 17:41:17.860175 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b301322e82e1a4c2a18d699e32b8eac5d3e309cc0928919cd9e29ca7b4e08b97"} err="failed to get container status \"b301322e82e1a4c2a18d699e32b8eac5d3e309cc0928919cd9e29ca7b4e08b97\": rpc error: code = NotFound desc = could not find container \"b301322e82e1a4c2a18d699e32b8eac5d3e309cc0928919cd9e29ca7b4e08b97\": container with ID starting with b301322e82e1a4c2a18d699e32b8eac5d3e309cc0928919cd9e29ca7b4e08b97 not found: ID does not exist" Dec 13 17:41:18 crc kubenswrapper[4989]: I1213 17:41:18.022125 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" path="/var/lib/kubelet/pods/03013fe5-bdd4-45a3-abec-1e32d6a017c1/volumes" Dec 13 17:41:27 crc kubenswrapper[4989]: I1213 17:41:27.854092 4989 scope.go:117] "RemoveContainer" containerID="cdcf3191a039b30957514bfec1acf1d206721c732f0d517924316f3ea2088c9b" Dec 13 17:41:27 crc kubenswrapper[4989]: I1213 17:41:27.871187 4989 scope.go:117] "RemoveContainer" containerID="411caefce67180e749b5a8c899d65ae03b2209262889a2e7ebd52d3d2831dc29" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.144995 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2fxz9/must-gather-xpxrr"] Dec 13 17:42:24 crc kubenswrapper[4989]: E1213 17:42:24.146528 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" containerName="registry-server" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.146602 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" containerName="registry-server" Dec 13 17:42:24 crc kubenswrapper[4989]: E1213 17:42:24.146663 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" containerName="extract-content" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.146715 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" containerName="extract-content" Dec 13 17:42:24 crc kubenswrapper[4989]: E1213 17:42:24.146780 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" containerName="extract-utilities" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.146859 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" containerName="extract-utilities" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.147007 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="03013fe5-bdd4-45a3-abec-1e32d6a017c1" containerName="registry-server" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.147657 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2fxz9/must-gather-xpxrr" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.150440 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-2fxz9"/"default-dockercfg-qwdv4" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.150661 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2fxz9"/"kube-root-ca.crt" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.150862 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2fxz9"/"openshift-service-ca.crt" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.164227 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2fxz9/must-gather-xpxrr"] Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.235416 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/385fa572-58e0-4f9f-a315-0b64863bd115-must-gather-output\") pod \"must-gather-xpxrr\" (UID: \"385fa572-58e0-4f9f-a315-0b64863bd115\") " pod="openshift-must-gather-2fxz9/must-gather-xpxrr" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.235474 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htlrl\" (UniqueName: \"kubernetes.io/projected/385fa572-58e0-4f9f-a315-0b64863bd115-kube-api-access-htlrl\") pod \"must-gather-xpxrr\" (UID: \"385fa572-58e0-4f9f-a315-0b64863bd115\") " pod="openshift-must-gather-2fxz9/must-gather-xpxrr" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.336376 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/385fa572-58e0-4f9f-a315-0b64863bd115-must-gather-output\") pod \"must-gather-xpxrr\" (UID: \"385fa572-58e0-4f9f-a315-0b64863bd115\") " pod="openshift-must-gather-2fxz9/must-gather-xpxrr" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.336450 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htlrl\" (UniqueName: \"kubernetes.io/projected/385fa572-58e0-4f9f-a315-0b64863bd115-kube-api-access-htlrl\") pod \"must-gather-xpxrr\" (UID: \"385fa572-58e0-4f9f-a315-0b64863bd115\") " pod="openshift-must-gather-2fxz9/must-gather-xpxrr" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.337000 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/385fa572-58e0-4f9f-a315-0b64863bd115-must-gather-output\") pod \"must-gather-xpxrr\" (UID: \"385fa572-58e0-4f9f-a315-0b64863bd115\") " pod="openshift-must-gather-2fxz9/must-gather-xpxrr" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.385384 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htlrl\" (UniqueName: \"kubernetes.io/projected/385fa572-58e0-4f9f-a315-0b64863bd115-kube-api-access-htlrl\") pod \"must-gather-xpxrr\" (UID: \"385fa572-58e0-4f9f-a315-0b64863bd115\") " pod="openshift-must-gather-2fxz9/must-gather-xpxrr" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.470540 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2fxz9/must-gather-xpxrr" Dec 13 17:42:24 crc kubenswrapper[4989]: I1213 17:42:24.711376 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2fxz9/must-gather-xpxrr"] Dec 13 17:42:25 crc kubenswrapper[4989]: I1213 17:42:25.245700 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2fxz9/must-gather-xpxrr" event={"ID":"385fa572-58e0-4f9f-a315-0b64863bd115","Type":"ContainerStarted","Data":"57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541"} Dec 13 17:42:25 crc kubenswrapper[4989]: I1213 17:42:25.245998 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2fxz9/must-gather-xpxrr" event={"ID":"385fa572-58e0-4f9f-a315-0b64863bd115","Type":"ContainerStarted","Data":"35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b"} Dec 13 17:42:25 crc kubenswrapper[4989]: I1213 17:42:25.246016 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2fxz9/must-gather-xpxrr" event={"ID":"385fa572-58e0-4f9f-a315-0b64863bd115","Type":"ContainerStarted","Data":"6d3529e16e949fad0bfddb2def0a678ffb6a720d89520209ff8517c5a87964eb"} Dec 13 17:42:25 crc kubenswrapper[4989]: I1213 17:42:25.261889 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2fxz9/must-gather-xpxrr" podStartSLOduration=1.261872826 podStartE2EDuration="1.261872826s" podCreationTimestamp="2025-12-13 17:42:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:42:25.259032737 +0000 UTC m=+1559.865479875" watchObservedRunningTime="2025-12-13 17:42:25.261872826 +0000 UTC m=+1559.868319964" Dec 13 17:42:27 crc kubenswrapper[4989]: I1213 17:42:27.951060 4989 scope.go:117] "RemoveContainer" containerID="8180eb9a5e8b9879718f47a629137ef23241a2b78ec10aa98a03edaa37cbe592" Dec 13 17:42:27 crc kubenswrapper[4989]: I1213 17:42:27.975811 4989 scope.go:117] "RemoveContainer" containerID="d6ebdd145cc977673fa30d15692a37f80f5a420dd27941c18ed7d3f959d1a388" Dec 13 17:42:28 crc kubenswrapper[4989]: I1213 17:42:28.003542 4989 scope.go:117] "RemoveContainer" containerID="ca86b6fae1da73c4c1ce6c36a602c4531e059363e24d326ec1d995e245f1a3ba" Dec 13 17:42:28 crc kubenswrapper[4989]: I1213 17:42:28.068630 4989 scope.go:117] "RemoveContainer" containerID="4fbe5371891617db8be6bf086c4bc30a0c16287339133510b5da3348cb525e90" Dec 13 17:42:28 crc kubenswrapper[4989]: I1213 17:42:28.081223 4989 scope.go:117] "RemoveContainer" containerID="a44f8e7ed939b3d1b2466733ef84a4104089db0d54dbff2c85a0cbe753d51b83" Dec 13 17:43:10 crc kubenswrapper[4989]: I1213 17:43:10.022193 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-fkvxp_a134b88a-8fa0-486b-b66d-b1df520f4a6c/control-plane-machine-set-operator/0.log" Dec 13 17:43:10 crc kubenswrapper[4989]: I1213 17:43:10.225452 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8hwgn_7fe4787f-cdd3-44fe-8405-dd6ebc9b455c/kube-rbac-proxy/0.log" Dec 13 17:43:10 crc kubenswrapper[4989]: I1213 17:43:10.282272 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8hwgn_7fe4787f-cdd3-44fe-8405-dd6ebc9b455c/machine-api-operator/0.log" Dec 13 17:43:26 crc kubenswrapper[4989]: I1213 17:43:26.206735 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-55pxp_c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29/kube-rbac-proxy/0.log" Dec 13 17:43:26 crc kubenswrapper[4989]: I1213 17:43:26.259024 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-55pxp_c3eeb7bd-0f35-4a83-8ac3-dec714e8ca29/controller/0.log" Dec 13 17:43:26 crc kubenswrapper[4989]: I1213 17:43:26.402215 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-frr-files/0.log" Dec 13 17:43:26 crc kubenswrapper[4989]: I1213 17:43:26.602988 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-frr-files/0.log" Dec 13 17:43:26 crc kubenswrapper[4989]: I1213 17:43:26.629418 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-reloader/0.log" Dec 13 17:43:26 crc kubenswrapper[4989]: I1213 17:43:26.642728 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-metrics/0.log" Dec 13 17:43:26 crc kubenswrapper[4989]: I1213 17:43:26.674773 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-reloader/0.log" Dec 13 17:43:26 crc kubenswrapper[4989]: I1213 17:43:26.829357 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-frr-files/0.log" Dec 13 17:43:26 crc kubenswrapper[4989]: I1213 17:43:26.862523 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-reloader/0.log" Dec 13 17:43:26 crc kubenswrapper[4989]: I1213 17:43:26.864208 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-metrics/0.log" Dec 13 17:43:26 crc kubenswrapper[4989]: I1213 17:43:26.921645 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-metrics/0.log" Dec 13 17:43:27 crc kubenswrapper[4989]: I1213 17:43:27.045447 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-frr-files/0.log" Dec 13 17:43:27 crc kubenswrapper[4989]: I1213 17:43:27.052446 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-reloader/0.log" Dec 13 17:43:27 crc kubenswrapper[4989]: I1213 17:43:27.074654 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/cp-metrics/0.log" Dec 13 17:43:27 crc kubenswrapper[4989]: I1213 17:43:27.132777 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/controller/0.log" Dec 13 17:43:27 crc kubenswrapper[4989]: I1213 17:43:27.241568 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/kube-rbac-proxy/0.log" Dec 13 17:43:27 crc kubenswrapper[4989]: I1213 17:43:27.247555 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/frr-metrics/0.log" Dec 13 17:43:27 crc kubenswrapper[4989]: I1213 17:43:27.337011 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/kube-rbac-proxy-frr/0.log" Dec 13 17:43:27 crc kubenswrapper[4989]: I1213 17:43:27.491369 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/reloader/0.log" Dec 13 17:43:27 crc kubenswrapper[4989]: I1213 17:43:27.578496 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-2shrx_7c4da269-7f3c-4101-8776-3c10b7379d9c/frr-k8s-webhook-server/0.log" Dec 13 17:43:27 crc kubenswrapper[4989]: I1213 17:43:27.670479 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hq8z9_f2005529-2d5f-4dc3-b783-d30e0e1d7962/frr/0.log" Dec 13 17:43:27 crc kubenswrapper[4989]: I1213 17:43:27.783143 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-64f9694c74-whn2q_1ba2b41f-9366-455a-ab5c-51a199ac8d52/manager/0.log" Dec 13 17:43:28 crc kubenswrapper[4989]: I1213 17:43:28.041829 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-84cdcb4857-zvzst_9f540b88-5e5f-4420-8b2c-d49f42087943/webhook-server/0.log" Dec 13 17:43:28 crc kubenswrapper[4989]: I1213 17:43:28.188068 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r5qp2_931ee02f-2887-4831-a0df-94eccdf81672/kube-rbac-proxy/0.log" Dec 13 17:43:28 crc kubenswrapper[4989]: I1213 17:43:28.193298 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-r5qp2_931ee02f-2887-4831-a0df-94eccdf81672/speaker/0.log" Dec 13 17:43:38 crc kubenswrapper[4989]: I1213 17:43:38.829309 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:43:38 crc kubenswrapper[4989]: I1213 17:43:38.829884 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.004544 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/util/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.216624 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/pull/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.232994 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/pull/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.234729 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/util/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.398522 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/util/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.403993 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/pull/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.426905 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d488kqc_d2d02100-c895-4311-955b-a06f5d7b1e0d/extract/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.566485 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-utilities/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.738907 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-content/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.740738 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-content/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.783646 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-utilities/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.938135 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-content/0.log" Dec 13 17:43:52 crc kubenswrapper[4989]: I1213 17:43:52.953269 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/extract-utilities/0.log" Dec 13 17:43:53 crc kubenswrapper[4989]: I1213 17:43:53.143457 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-utilities/0.log" Dec 13 17:43:53 crc kubenswrapper[4989]: I1213 17:43:53.333562 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-content/0.log" Dec 13 17:43:53 crc kubenswrapper[4989]: I1213 17:43:53.343434 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-content/0.log" Dec 13 17:43:53 crc kubenswrapper[4989]: I1213 17:43:53.369916 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r59g9_bc88ed3d-34b8-4968-a381-9bf45fe374c1/registry-server/0.log" Dec 13 17:43:53 crc kubenswrapper[4989]: I1213 17:43:53.428153 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-utilities/0.log" Dec 13 17:43:53 crc kubenswrapper[4989]: I1213 17:43:53.531205 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-utilities/0.log" Dec 13 17:43:53 crc kubenswrapper[4989]: I1213 17:43:53.558503 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/extract-content/0.log" Dec 13 17:43:53 crc kubenswrapper[4989]: I1213 17:43:53.797977 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wl78k_a56dab99-6f88-4da4-af60-913e0b4c3be6/marketplace-operator/0.log" Dec 13 17:43:53 crc kubenswrapper[4989]: I1213 17:43:53.802839 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-utilities/0.log" Dec 13 17:43:53 crc kubenswrapper[4989]: I1213 17:43:53.908974 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-6bhwp_981677fb-9915-45ba-83c3-bb602abaf7d4/registry-server/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.028352 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-content/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.029979 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-content/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.073054 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-utilities/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.245461 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-content/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.252192 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/extract-utilities/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.370229 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g542s_98203275-48dd-4bcd-a4c0-709142c9e024/registry-server/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.405076 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-utilities/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.616242 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-content/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.622637 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-utilities/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.638656 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-content/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.806313 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-utilities/0.log" Dec 13 17:43:54 crc kubenswrapper[4989]: I1213 17:43:54.813256 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/extract-content/0.log" Dec 13 17:43:55 crc kubenswrapper[4989]: I1213 17:43:55.043814 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xsmj4_1e515c25-23f4-49f2-bc3f-0fbb5e8046be/registry-server/0.log" Dec 13 17:44:08 crc kubenswrapper[4989]: I1213 17:44:08.829638 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:44:08 crc kubenswrapper[4989]: I1213 17:44:08.830284 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:44:38 crc kubenswrapper[4989]: I1213 17:44:38.829153 4989 patch_prober.go:28] interesting pod/machine-config-daemon-nh9k2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 13 17:44:38 crc kubenswrapper[4989]: I1213 17:44:38.829611 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 13 17:44:38 crc kubenswrapper[4989]: I1213 17:44:38.829648 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" Dec 13 17:44:38 crc kubenswrapper[4989]: I1213 17:44:38.830103 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2"} pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 13 17:44:38 crc kubenswrapper[4989]: I1213 17:44:38.830143 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerName="machine-config-daemon" containerID="cri-o://24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" gracePeriod=600 Dec 13 17:44:38 crc kubenswrapper[4989]: E1213 17:44:38.949981 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:44:38 crc kubenswrapper[4989]: I1213 17:44:38.963701 4989 generic.go:334] "Generic (PLEG): container finished" podID="a2b01148-171a-4f86-84a7-d326739e0dcf" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" exitCode=0 Dec 13 17:44:38 crc kubenswrapper[4989]: I1213 17:44:38.963780 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" event={"ID":"a2b01148-171a-4f86-84a7-d326739e0dcf","Type":"ContainerDied","Data":"24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2"} Dec 13 17:44:38 crc kubenswrapper[4989]: I1213 17:44:38.963870 4989 scope.go:117] "RemoveContainer" containerID="0b5b47b60f26e4730dadda3c98b1d3052f3ae310472ca5f800e8c41049d96126" Dec 13 17:44:38 crc kubenswrapper[4989]: I1213 17:44:38.964543 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:44:38 crc kubenswrapper[4989]: E1213 17:44:38.964835 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:44:53 crc kubenswrapper[4989]: I1213 17:44:53.013978 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:44:53 crc kubenswrapper[4989]: E1213 17:44:53.016932 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.153782 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4"] Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.159018 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.165196 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.165312 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.171450 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4"] Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.272118 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09fb1cc4-f5f4-481c-9704-14c48c49cb28-secret-volume\") pod \"collect-profiles-29427465-vhpp4\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.272161 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09fb1cc4-f5f4-481c-9704-14c48c49cb28-config-volume\") pod \"collect-profiles-29427465-vhpp4\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.272220 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92pcc\" (UniqueName: \"kubernetes.io/projected/09fb1cc4-f5f4-481c-9704-14c48c49cb28-kube-api-access-92pcc\") pod \"collect-profiles-29427465-vhpp4\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.373096 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09fb1cc4-f5f4-481c-9704-14c48c49cb28-secret-volume\") pod \"collect-profiles-29427465-vhpp4\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.373138 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09fb1cc4-f5f4-481c-9704-14c48c49cb28-config-volume\") pod \"collect-profiles-29427465-vhpp4\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.373163 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92pcc\" (UniqueName: \"kubernetes.io/projected/09fb1cc4-f5f4-481c-9704-14c48c49cb28-kube-api-access-92pcc\") pod \"collect-profiles-29427465-vhpp4\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.375108 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09fb1cc4-f5f4-481c-9704-14c48c49cb28-config-volume\") pod \"collect-profiles-29427465-vhpp4\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.390167 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09fb1cc4-f5f4-481c-9704-14c48c49cb28-secret-volume\") pod \"collect-profiles-29427465-vhpp4\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.392091 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92pcc\" (UniqueName: \"kubernetes.io/projected/09fb1cc4-f5f4-481c-9704-14c48c49cb28-kube-api-access-92pcc\") pod \"collect-profiles-29427465-vhpp4\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.494881 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:00 crc kubenswrapper[4989]: I1213 17:45:00.908416 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4"] Dec 13 17:45:01 crc kubenswrapper[4989]: I1213 17:45:01.125963 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" event={"ID":"09fb1cc4-f5f4-481c-9704-14c48c49cb28","Type":"ContainerStarted","Data":"0ac1838c49f9041152a340b73d49e89af8e242c5b46a1a255733225c8cc830c9"} Dec 13 17:45:01 crc kubenswrapper[4989]: I1213 17:45:01.126351 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" event={"ID":"09fb1cc4-f5f4-481c-9704-14c48c49cb28","Type":"ContainerStarted","Data":"a993a759822f418973efc048f9119697a3bd5fe0a4c61665b7106a1cd985836c"} Dec 13 17:45:01 crc kubenswrapper[4989]: I1213 17:45:01.150358 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" podStartSLOduration=1.150339505 podStartE2EDuration="1.150339505s" podCreationTimestamp="2025-12-13 17:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-13 17:45:01.148244169 +0000 UTC m=+1715.754691307" watchObservedRunningTime="2025-12-13 17:45:01.150339505 +0000 UTC m=+1715.756786643" Dec 13 17:45:02 crc kubenswrapper[4989]: I1213 17:45:02.132836 4989 generic.go:334] "Generic (PLEG): container finished" podID="09fb1cc4-f5f4-481c-9704-14c48c49cb28" containerID="0ac1838c49f9041152a340b73d49e89af8e242c5b46a1a255733225c8cc830c9" exitCode=0 Dec 13 17:45:02 crc kubenswrapper[4989]: I1213 17:45:02.132878 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" event={"ID":"09fb1cc4-f5f4-481c-9704-14c48c49cb28","Type":"ContainerDied","Data":"0ac1838c49f9041152a340b73d49e89af8e242c5b46a1a255733225c8cc830c9"} Dec 13 17:45:03 crc kubenswrapper[4989]: I1213 17:45:03.394142 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:03 crc kubenswrapper[4989]: I1213 17:45:03.519599 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09fb1cc4-f5f4-481c-9704-14c48c49cb28-config-volume\") pod \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " Dec 13 17:45:03 crc kubenswrapper[4989]: I1213 17:45:03.519645 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92pcc\" (UniqueName: \"kubernetes.io/projected/09fb1cc4-f5f4-481c-9704-14c48c49cb28-kube-api-access-92pcc\") pod \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " Dec 13 17:45:03 crc kubenswrapper[4989]: I1213 17:45:03.519835 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09fb1cc4-f5f4-481c-9704-14c48c49cb28-secret-volume\") pod \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\" (UID: \"09fb1cc4-f5f4-481c-9704-14c48c49cb28\") " Dec 13 17:45:03 crc kubenswrapper[4989]: I1213 17:45:03.520100 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09fb1cc4-f5f4-481c-9704-14c48c49cb28-config-volume" (OuterVolumeSpecName: "config-volume") pod "09fb1cc4-f5f4-481c-9704-14c48c49cb28" (UID: "09fb1cc4-f5f4-481c-9704-14c48c49cb28"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 17:45:03 crc kubenswrapper[4989]: I1213 17:45:03.527630 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09fb1cc4-f5f4-481c-9704-14c48c49cb28-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "09fb1cc4-f5f4-481c-9704-14c48c49cb28" (UID: "09fb1cc4-f5f4-481c-9704-14c48c49cb28"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 17:45:03 crc kubenswrapper[4989]: I1213 17:45:03.532261 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09fb1cc4-f5f4-481c-9704-14c48c49cb28-kube-api-access-92pcc" (OuterVolumeSpecName: "kube-api-access-92pcc") pod "09fb1cc4-f5f4-481c-9704-14c48c49cb28" (UID: "09fb1cc4-f5f4-481c-9704-14c48c49cb28"). InnerVolumeSpecName "kube-api-access-92pcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:45:03 crc kubenswrapper[4989]: I1213 17:45:03.621667 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09fb1cc4-f5f4-481c-9704-14c48c49cb28-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 13 17:45:03 crc kubenswrapper[4989]: I1213 17:45:03.621714 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09fb1cc4-f5f4-481c-9704-14c48c49cb28-config-volume\") on node \"crc\" DevicePath \"\"" Dec 13 17:45:03 crc kubenswrapper[4989]: I1213 17:45:03.621726 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92pcc\" (UniqueName: \"kubernetes.io/projected/09fb1cc4-f5f4-481c-9704-14c48c49cb28-kube-api-access-92pcc\") on node \"crc\" DevicePath \"\"" Dec 13 17:45:04 crc kubenswrapper[4989]: I1213 17:45:04.144070 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" event={"ID":"09fb1cc4-f5f4-481c-9704-14c48c49cb28","Type":"ContainerDied","Data":"a993a759822f418973efc048f9119697a3bd5fe0a4c61665b7106a1cd985836c"} Dec 13 17:45:04 crc kubenswrapper[4989]: I1213 17:45:04.144107 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a993a759822f418973efc048f9119697a3bd5fe0a4c61665b7106a1cd985836c" Dec 13 17:45:04 crc kubenswrapper[4989]: I1213 17:45:04.144159 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29427465-vhpp4" Dec 13 17:45:05 crc kubenswrapper[4989]: I1213 17:45:05.155335 4989 generic.go:334] "Generic (PLEG): container finished" podID="385fa572-58e0-4f9f-a315-0b64863bd115" containerID="35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b" exitCode=0 Dec 13 17:45:05 crc kubenswrapper[4989]: I1213 17:45:05.155422 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2fxz9/must-gather-xpxrr" event={"ID":"385fa572-58e0-4f9f-a315-0b64863bd115","Type":"ContainerDied","Data":"35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b"} Dec 13 17:45:05 crc kubenswrapper[4989]: I1213 17:45:05.156049 4989 scope.go:117] "RemoveContainer" containerID="35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b" Dec 13 17:45:05 crc kubenswrapper[4989]: I1213 17:45:05.346854 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2fxz9_must-gather-xpxrr_385fa572-58e0-4f9f-a315-0b64863bd115/gather/0.log" Dec 13 17:45:07 crc kubenswrapper[4989]: I1213 17:45:07.014986 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:45:07 crc kubenswrapper[4989]: E1213 17:45:07.015448 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:45:14 crc kubenswrapper[4989]: I1213 17:45:14.178205 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2fxz9/must-gather-xpxrr"] Dec 13 17:45:14 crc kubenswrapper[4989]: I1213 17:45:14.178962 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-2fxz9/must-gather-xpxrr" podUID="385fa572-58e0-4f9f-a315-0b64863bd115" containerName="copy" containerID="cri-o://57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541" gracePeriod=2 Dec 13 17:45:14 crc kubenswrapper[4989]: I1213 17:45:14.184930 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2fxz9/must-gather-xpxrr"] Dec 13 17:45:14 crc kubenswrapper[4989]: I1213 17:45:14.556921 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2fxz9_must-gather-xpxrr_385fa572-58e0-4f9f-a315-0b64863bd115/copy/0.log" Dec 13 17:45:14 crc kubenswrapper[4989]: I1213 17:45:14.557275 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2fxz9/must-gather-xpxrr" Dec 13 17:45:14 crc kubenswrapper[4989]: I1213 17:45:14.692479 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htlrl\" (UniqueName: \"kubernetes.io/projected/385fa572-58e0-4f9f-a315-0b64863bd115-kube-api-access-htlrl\") pod \"385fa572-58e0-4f9f-a315-0b64863bd115\" (UID: \"385fa572-58e0-4f9f-a315-0b64863bd115\") " Dec 13 17:45:14 crc kubenswrapper[4989]: I1213 17:45:14.692564 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/385fa572-58e0-4f9f-a315-0b64863bd115-must-gather-output\") pod \"385fa572-58e0-4f9f-a315-0b64863bd115\" (UID: \"385fa572-58e0-4f9f-a315-0b64863bd115\") " Dec 13 17:45:14 crc kubenswrapper[4989]: I1213 17:45:14.705988 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/385fa572-58e0-4f9f-a315-0b64863bd115-kube-api-access-htlrl" (OuterVolumeSpecName: "kube-api-access-htlrl") pod "385fa572-58e0-4f9f-a315-0b64863bd115" (UID: "385fa572-58e0-4f9f-a315-0b64863bd115"). InnerVolumeSpecName "kube-api-access-htlrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:45:14 crc kubenswrapper[4989]: I1213 17:45:14.750957 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/385fa572-58e0-4f9f-a315-0b64863bd115-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "385fa572-58e0-4f9f-a315-0b64863bd115" (UID: "385fa572-58e0-4f9f-a315-0b64863bd115"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:45:14 crc kubenswrapper[4989]: I1213 17:45:14.793736 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htlrl\" (UniqueName: \"kubernetes.io/projected/385fa572-58e0-4f9f-a315-0b64863bd115-kube-api-access-htlrl\") on node \"crc\" DevicePath \"\"" Dec 13 17:45:14 crc kubenswrapper[4989]: I1213 17:45:14.793763 4989 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/385fa572-58e0-4f9f-a315-0b64863bd115-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 13 17:45:15 crc kubenswrapper[4989]: I1213 17:45:15.216398 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2fxz9_must-gather-xpxrr_385fa572-58e0-4f9f-a315-0b64863bd115/copy/0.log" Dec 13 17:45:15 crc kubenswrapper[4989]: I1213 17:45:15.217777 4989 generic.go:334] "Generic (PLEG): container finished" podID="385fa572-58e0-4f9f-a315-0b64863bd115" containerID="57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541" exitCode=143 Dec 13 17:45:15 crc kubenswrapper[4989]: I1213 17:45:15.217937 4989 scope.go:117] "RemoveContainer" containerID="57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541" Dec 13 17:45:15 crc kubenswrapper[4989]: I1213 17:45:15.218226 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2fxz9/must-gather-xpxrr" Dec 13 17:45:15 crc kubenswrapper[4989]: I1213 17:45:15.245607 4989 scope.go:117] "RemoveContainer" containerID="35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b" Dec 13 17:45:15 crc kubenswrapper[4989]: I1213 17:45:15.297684 4989 scope.go:117] "RemoveContainer" containerID="57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541" Dec 13 17:45:15 crc kubenswrapper[4989]: E1213 17:45:15.298327 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541\": container with ID starting with 57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541 not found: ID does not exist" containerID="57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541" Dec 13 17:45:15 crc kubenswrapper[4989]: I1213 17:45:15.298388 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541"} err="failed to get container status \"57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541\": rpc error: code = NotFound desc = could not find container \"57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541\": container with ID starting with 57878d81a2bbe0fa0b4616ddba5593b9f486613a1aff49de3c923b65a98e2541 not found: ID does not exist" Dec 13 17:45:15 crc kubenswrapper[4989]: I1213 17:45:15.298432 4989 scope.go:117] "RemoveContainer" containerID="35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b" Dec 13 17:45:15 crc kubenswrapper[4989]: E1213 17:45:15.299010 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b\": container with ID starting with 35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b not found: ID does not exist" containerID="35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b" Dec 13 17:45:15 crc kubenswrapper[4989]: I1213 17:45:15.299046 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b"} err="failed to get container status \"35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b\": rpc error: code = NotFound desc = could not find container \"35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b\": container with ID starting with 35eb7f1a32eac0da8d22d8dc9f5ed6fb03ff65be83401b837232dca25d60426b not found: ID does not exist" Dec 13 17:45:16 crc kubenswrapper[4989]: I1213 17:45:16.025142 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="385fa572-58e0-4f9f-a315-0b64863bd115" path="/var/lib/kubelet/pods/385fa572-58e0-4f9f-a315-0b64863bd115/volumes" Dec 13 17:45:19 crc kubenswrapper[4989]: I1213 17:45:19.014393 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:45:19 crc kubenswrapper[4989]: E1213 17:45:19.015132 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:45:31 crc kubenswrapper[4989]: I1213 17:45:31.014681 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:45:31 crc kubenswrapper[4989]: E1213 17:45:31.015585 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:45:45 crc kubenswrapper[4989]: I1213 17:45:45.015327 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:45:45 crc kubenswrapper[4989]: E1213 17:45:45.016221 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:45:59 crc kubenswrapper[4989]: I1213 17:45:59.015716 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:45:59 crc kubenswrapper[4989]: E1213 17:45:59.016328 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:46:13 crc kubenswrapper[4989]: I1213 17:46:13.014296 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:46:13 crc kubenswrapper[4989]: E1213 17:46:13.015323 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:46:24 crc kubenswrapper[4989]: I1213 17:46:24.014386 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:46:24 crc kubenswrapper[4989]: E1213 17:46:24.015676 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:46:36 crc kubenswrapper[4989]: I1213 17:46:36.016866 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:46:36 crc kubenswrapper[4989]: E1213 17:46:36.021766 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:46:49 crc kubenswrapper[4989]: I1213 17:46:49.014887 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:46:49 crc kubenswrapper[4989]: E1213 17:46:49.016358 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:47:02 crc kubenswrapper[4989]: I1213 17:47:02.014679 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:47:02 crc kubenswrapper[4989]: E1213 17:47:02.015429 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.851846 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-25cb7"] Dec 13 17:47:13 crc kubenswrapper[4989]: E1213 17:47:13.852837 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385fa572-58e0-4f9f-a315-0b64863bd115" containerName="copy" Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.852855 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="385fa572-58e0-4f9f-a315-0b64863bd115" containerName="copy" Dec 13 17:47:13 crc kubenswrapper[4989]: E1213 17:47:13.852886 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385fa572-58e0-4f9f-a315-0b64863bd115" containerName="gather" Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.852895 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="385fa572-58e0-4f9f-a315-0b64863bd115" containerName="gather" Dec 13 17:47:13 crc kubenswrapper[4989]: E1213 17:47:13.852909 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09fb1cc4-f5f4-481c-9704-14c48c49cb28" containerName="collect-profiles" Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.852918 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="09fb1cc4-f5f4-481c-9704-14c48c49cb28" containerName="collect-profiles" Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.853069 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="385fa572-58e0-4f9f-a315-0b64863bd115" containerName="gather" Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.853087 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="09fb1cc4-f5f4-481c-9704-14c48c49cb28" containerName="collect-profiles" Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.853098 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="385fa572-58e0-4f9f-a315-0b64863bd115" containerName="copy" Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.854066 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.863782 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-25cb7"] Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.984424 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-utilities\") pod \"community-operators-25cb7\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.984476 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqrtw\" (UniqueName: \"kubernetes.io/projected/76539ec9-6f61-40ec-baa7-59176c4bc809-kube-api-access-dqrtw\") pod \"community-operators-25cb7\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:13 crc kubenswrapper[4989]: I1213 17:47:13.984581 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-catalog-content\") pod \"community-operators-25cb7\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.086141 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-catalog-content\") pod \"community-operators-25cb7\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.086212 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-utilities\") pod \"community-operators-25cb7\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.086239 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqrtw\" (UniqueName: \"kubernetes.io/projected/76539ec9-6f61-40ec-baa7-59176c4bc809-kube-api-access-dqrtw\") pod \"community-operators-25cb7\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.086961 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-catalog-content\") pod \"community-operators-25cb7\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.086969 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-utilities\") pod \"community-operators-25cb7\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.106958 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqrtw\" (UniqueName: \"kubernetes.io/projected/76539ec9-6f61-40ec-baa7-59176c4bc809-kube-api-access-dqrtw\") pod \"community-operators-25cb7\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.200182 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.456659 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-25cb7"] Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.948261 4989 generic.go:334] "Generic (PLEG): container finished" podID="76539ec9-6f61-40ec-baa7-59176c4bc809" containerID="faaffbbb0e5c94741a47834dbd227efdcd85636ef8ad309bc8eef1b5b72871e8" exitCode=0 Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.948301 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-25cb7" event={"ID":"76539ec9-6f61-40ec-baa7-59176c4bc809","Type":"ContainerDied","Data":"faaffbbb0e5c94741a47834dbd227efdcd85636ef8ad309bc8eef1b5b72871e8"} Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.948327 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-25cb7" event={"ID":"76539ec9-6f61-40ec-baa7-59176c4bc809","Type":"ContainerStarted","Data":"cdc39add611a1c838d0657cf1f6db235cc9265ef90f51f95f462fbf8bacff95b"} Dec 13 17:47:14 crc kubenswrapper[4989]: I1213 17:47:14.949995 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 13 17:47:15 crc kubenswrapper[4989]: I1213 17:47:15.954656 4989 generic.go:334] "Generic (PLEG): container finished" podID="76539ec9-6f61-40ec-baa7-59176c4bc809" containerID="5e2566c806fc16a98ef48ce7ed7c8685f96d837069a3561f6f5b989e39e26b3e" exitCode=0 Dec 13 17:47:15 crc kubenswrapper[4989]: I1213 17:47:15.954761 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-25cb7" event={"ID":"76539ec9-6f61-40ec-baa7-59176c4bc809","Type":"ContainerDied","Data":"5e2566c806fc16a98ef48ce7ed7c8685f96d837069a3561f6f5b989e39e26b3e"} Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.017287 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:47:16 crc kubenswrapper[4989]: E1213 17:47:16.017536 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.254118 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r9frb"] Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.256885 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.259703 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r9frb"] Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.318701 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-catalog-content\") pod \"redhat-operators-r9frb\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.318740 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mc4x\" (UniqueName: \"kubernetes.io/projected/23748c76-d92f-46cd-8300-7e80230e3241-kube-api-access-4mc4x\") pod \"redhat-operators-r9frb\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.318807 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-utilities\") pod \"redhat-operators-r9frb\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.420774 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-utilities\") pod \"redhat-operators-r9frb\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.420904 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-catalog-content\") pod \"redhat-operators-r9frb\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.420936 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mc4x\" (UniqueName: \"kubernetes.io/projected/23748c76-d92f-46cd-8300-7e80230e3241-kube-api-access-4mc4x\") pod \"redhat-operators-r9frb\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.421978 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-utilities\") pod \"redhat-operators-r9frb\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.422408 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-catalog-content\") pod \"redhat-operators-r9frb\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.450066 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mc4x\" (UniqueName: \"kubernetes.io/projected/23748c76-d92f-46cd-8300-7e80230e3241-kube-api-access-4mc4x\") pod \"redhat-operators-r9frb\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.603509 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.817638 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r9frb"] Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.963278 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-25cb7" event={"ID":"76539ec9-6f61-40ec-baa7-59176c4bc809","Type":"ContainerStarted","Data":"72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35"} Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.964096 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9frb" event={"ID":"23748c76-d92f-46cd-8300-7e80230e3241","Type":"ContainerStarted","Data":"8754205debe8fccbdb801d6069546d688ad404065775fcd14eb5948c4a6bc3ca"} Dec 13 17:47:16 crc kubenswrapper[4989]: I1213 17:47:16.989337 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-25cb7" podStartSLOduration=2.561279238 podStartE2EDuration="3.989317033s" podCreationTimestamp="2025-12-13 17:47:13 +0000 UTC" firstStartedPulling="2025-12-13 17:47:14.94971751 +0000 UTC m=+1849.556164648" lastFinishedPulling="2025-12-13 17:47:16.377755305 +0000 UTC m=+1850.984202443" observedRunningTime="2025-12-13 17:47:16.985025799 +0000 UTC m=+1851.591472947" watchObservedRunningTime="2025-12-13 17:47:16.989317033 +0000 UTC m=+1851.595764171" Dec 13 17:47:17 crc kubenswrapper[4989]: I1213 17:47:17.970987 4989 generic.go:334] "Generic (PLEG): container finished" podID="23748c76-d92f-46cd-8300-7e80230e3241" containerID="1025a3f12b3cf8e147b74db5867e86eba9d62d1ba7e279349b41c2a5dbaf4a13" exitCode=0 Dec 13 17:47:17 crc kubenswrapper[4989]: I1213 17:47:17.971178 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9frb" event={"ID":"23748c76-d92f-46cd-8300-7e80230e3241","Type":"ContainerDied","Data":"1025a3f12b3cf8e147b74db5867e86eba9d62d1ba7e279349b41c2a5dbaf4a13"} Dec 13 17:47:18 crc kubenswrapper[4989]: I1213 17:47:18.981063 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9frb" event={"ID":"23748c76-d92f-46cd-8300-7e80230e3241","Type":"ContainerStarted","Data":"29a53032f5c82b4787199a306b498f63dc758f87ee690339d8ed5c506db68179"} Dec 13 17:47:19 crc kubenswrapper[4989]: I1213 17:47:19.991648 4989 generic.go:334] "Generic (PLEG): container finished" podID="23748c76-d92f-46cd-8300-7e80230e3241" containerID="29a53032f5c82b4787199a306b498f63dc758f87ee690339d8ed5c506db68179" exitCode=0 Dec 13 17:47:19 crc kubenswrapper[4989]: I1213 17:47:19.991769 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9frb" event={"ID":"23748c76-d92f-46cd-8300-7e80230e3241","Type":"ContainerDied","Data":"29a53032f5c82b4787199a306b498f63dc758f87ee690339d8ed5c506db68179"} Dec 13 17:47:21 crc kubenswrapper[4989]: I1213 17:47:20.999922 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9frb" event={"ID":"23748c76-d92f-46cd-8300-7e80230e3241","Type":"ContainerStarted","Data":"a096ac635950f3a1d015090b3c2c8e921ed26d9c9d68e164bb4a278cf5b2bda7"} Dec 13 17:47:24 crc kubenswrapper[4989]: I1213 17:47:24.201523 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:24 crc kubenswrapper[4989]: I1213 17:47:24.201647 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:24 crc kubenswrapper[4989]: I1213 17:47:24.251511 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:24 crc kubenswrapper[4989]: I1213 17:47:24.286384 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r9frb" podStartSLOduration=5.810680648 podStartE2EDuration="8.286356016s" podCreationTimestamp="2025-12-13 17:47:16 +0000 UTC" firstStartedPulling="2025-12-13 17:47:17.97255242 +0000 UTC m=+1852.578999558" lastFinishedPulling="2025-12-13 17:47:20.448227798 +0000 UTC m=+1855.054674926" observedRunningTime="2025-12-13 17:47:21.021731918 +0000 UTC m=+1855.628179056" watchObservedRunningTime="2025-12-13 17:47:24.286356016 +0000 UTC m=+1858.892803154" Dec 13 17:47:25 crc kubenswrapper[4989]: I1213 17:47:25.093029 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:25 crc kubenswrapper[4989]: I1213 17:47:25.146284 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-25cb7"] Dec 13 17:47:26 crc kubenswrapper[4989]: I1213 17:47:26.604032 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:26 crc kubenswrapper[4989]: I1213 17:47:26.604075 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:26 crc kubenswrapper[4989]: I1213 17:47:26.661620 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:27 crc kubenswrapper[4989]: I1213 17:47:27.032684 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-25cb7" podUID="76539ec9-6f61-40ec-baa7-59176c4bc809" containerName="registry-server" containerID="cri-o://72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35" gracePeriod=2 Dec 13 17:47:27 crc kubenswrapper[4989]: I1213 17:47:27.083240 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:27 crc kubenswrapper[4989]: I1213 17:47:27.444073 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r9frb"] Dec 13 17:47:27 crc kubenswrapper[4989]: I1213 17:47:27.890641 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:27 crc kubenswrapper[4989]: I1213 17:47:27.991131 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqrtw\" (UniqueName: \"kubernetes.io/projected/76539ec9-6f61-40ec-baa7-59176c4bc809-kube-api-access-dqrtw\") pod \"76539ec9-6f61-40ec-baa7-59176c4bc809\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " Dec 13 17:47:27 crc kubenswrapper[4989]: I1213 17:47:27.991198 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-utilities\") pod \"76539ec9-6f61-40ec-baa7-59176c4bc809\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " Dec 13 17:47:27 crc kubenswrapper[4989]: I1213 17:47:27.991340 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-catalog-content\") pod \"76539ec9-6f61-40ec-baa7-59176c4bc809\" (UID: \"76539ec9-6f61-40ec-baa7-59176c4bc809\") " Dec 13 17:47:27 crc kubenswrapper[4989]: I1213 17:47:27.992250 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-utilities" (OuterVolumeSpecName: "utilities") pod "76539ec9-6f61-40ec-baa7-59176c4bc809" (UID: "76539ec9-6f61-40ec-baa7-59176c4bc809"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.000996 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76539ec9-6f61-40ec-baa7-59176c4bc809-kube-api-access-dqrtw" (OuterVolumeSpecName: "kube-api-access-dqrtw") pod "76539ec9-6f61-40ec-baa7-59176c4bc809" (UID: "76539ec9-6f61-40ec-baa7-59176c4bc809"). InnerVolumeSpecName "kube-api-access-dqrtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.046943 4989 generic.go:334] "Generic (PLEG): container finished" podID="76539ec9-6f61-40ec-baa7-59176c4bc809" containerID="72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35" exitCode=0 Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.047003 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-25cb7" event={"ID":"76539ec9-6f61-40ec-baa7-59176c4bc809","Type":"ContainerDied","Data":"72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35"} Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.047053 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-25cb7" event={"ID":"76539ec9-6f61-40ec-baa7-59176c4bc809","Type":"ContainerDied","Data":"cdc39add611a1c838d0657cf1f6db235cc9265ef90f51f95f462fbf8bacff95b"} Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.047069 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-25cb7" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.047134 4989 scope.go:117] "RemoveContainer" containerID="72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.060331 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76539ec9-6f61-40ec-baa7-59176c4bc809" (UID: "76539ec9-6f61-40ec-baa7-59176c4bc809"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.074293 4989 scope.go:117] "RemoveContainer" containerID="5e2566c806fc16a98ef48ce7ed7c8685f96d837069a3561f6f5b989e39e26b3e" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.092359 4989 scope.go:117] "RemoveContainer" containerID="faaffbbb0e5c94741a47834dbd227efdcd85636ef8ad309bc8eef1b5b72871e8" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.092762 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.092820 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqrtw\" (UniqueName: \"kubernetes.io/projected/76539ec9-6f61-40ec-baa7-59176c4bc809-kube-api-access-dqrtw\") on node \"crc\" DevicePath \"\"" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.092834 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76539ec9-6f61-40ec-baa7-59176c4bc809-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.112362 4989 scope.go:117] "RemoveContainer" containerID="72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35" Dec 13 17:47:28 crc kubenswrapper[4989]: E1213 17:47:28.115135 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35\": container with ID starting with 72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35 not found: ID does not exist" containerID="72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.115204 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35"} err="failed to get container status \"72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35\": rpc error: code = NotFound desc = could not find container \"72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35\": container with ID starting with 72a655cb36ebc38c7bb6cef4c647cf55a6d1443078026146e22ad3eae4466a35 not found: ID does not exist" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.115260 4989 scope.go:117] "RemoveContainer" containerID="5e2566c806fc16a98ef48ce7ed7c8685f96d837069a3561f6f5b989e39e26b3e" Dec 13 17:47:28 crc kubenswrapper[4989]: E1213 17:47:28.116370 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e2566c806fc16a98ef48ce7ed7c8685f96d837069a3561f6f5b989e39e26b3e\": container with ID starting with 5e2566c806fc16a98ef48ce7ed7c8685f96d837069a3561f6f5b989e39e26b3e not found: ID does not exist" containerID="5e2566c806fc16a98ef48ce7ed7c8685f96d837069a3561f6f5b989e39e26b3e" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.116400 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e2566c806fc16a98ef48ce7ed7c8685f96d837069a3561f6f5b989e39e26b3e"} err="failed to get container status \"5e2566c806fc16a98ef48ce7ed7c8685f96d837069a3561f6f5b989e39e26b3e\": rpc error: code = NotFound desc = could not find container \"5e2566c806fc16a98ef48ce7ed7c8685f96d837069a3561f6f5b989e39e26b3e\": container with ID starting with 5e2566c806fc16a98ef48ce7ed7c8685f96d837069a3561f6f5b989e39e26b3e not found: ID does not exist" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.116422 4989 scope.go:117] "RemoveContainer" containerID="faaffbbb0e5c94741a47834dbd227efdcd85636ef8ad309bc8eef1b5b72871e8" Dec 13 17:47:28 crc kubenswrapper[4989]: E1213 17:47:28.117699 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faaffbbb0e5c94741a47834dbd227efdcd85636ef8ad309bc8eef1b5b72871e8\": container with ID starting with faaffbbb0e5c94741a47834dbd227efdcd85636ef8ad309bc8eef1b5b72871e8 not found: ID does not exist" containerID="faaffbbb0e5c94741a47834dbd227efdcd85636ef8ad309bc8eef1b5b72871e8" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.117752 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faaffbbb0e5c94741a47834dbd227efdcd85636ef8ad309bc8eef1b5b72871e8"} err="failed to get container status \"faaffbbb0e5c94741a47834dbd227efdcd85636ef8ad309bc8eef1b5b72871e8\": rpc error: code = NotFound desc = could not find container \"faaffbbb0e5c94741a47834dbd227efdcd85636ef8ad309bc8eef1b5b72871e8\": container with ID starting with faaffbbb0e5c94741a47834dbd227efdcd85636ef8ad309bc8eef1b5b72871e8 not found: ID does not exist" Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.375984 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-25cb7"] Dec 13 17:47:28 crc kubenswrapper[4989]: I1213 17:47:28.379136 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-25cb7"] Dec 13 17:47:29 crc kubenswrapper[4989]: I1213 17:47:29.014735 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:47:29 crc kubenswrapper[4989]: E1213 17:47:29.015846 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:47:29 crc kubenswrapper[4989]: I1213 17:47:29.057524 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r9frb" podUID="23748c76-d92f-46cd-8300-7e80230e3241" containerName="registry-server" containerID="cri-o://a096ac635950f3a1d015090b3c2c8e921ed26d9c9d68e164bb4a278cf5b2bda7" gracePeriod=2 Dec 13 17:47:30 crc kubenswrapper[4989]: I1213 17:47:30.022953 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76539ec9-6f61-40ec-baa7-59176c4bc809" path="/var/lib/kubelet/pods/76539ec9-6f61-40ec-baa7-59176c4bc809/volumes" Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.072545 4989 generic.go:334] "Generic (PLEG): container finished" podID="23748c76-d92f-46cd-8300-7e80230e3241" containerID="a096ac635950f3a1d015090b3c2c8e921ed26d9c9d68e164bb4a278cf5b2bda7" exitCode=0 Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.072589 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9frb" event={"ID":"23748c76-d92f-46cd-8300-7e80230e3241","Type":"ContainerDied","Data":"a096ac635950f3a1d015090b3c2c8e921ed26d9c9d68e164bb4a278cf5b2bda7"} Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.226771 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.254373 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mc4x\" (UniqueName: \"kubernetes.io/projected/23748c76-d92f-46cd-8300-7e80230e3241-kube-api-access-4mc4x\") pod \"23748c76-d92f-46cd-8300-7e80230e3241\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.254441 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-utilities\") pod \"23748c76-d92f-46cd-8300-7e80230e3241\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.254482 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-catalog-content\") pod \"23748c76-d92f-46cd-8300-7e80230e3241\" (UID: \"23748c76-d92f-46cd-8300-7e80230e3241\") " Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.255450 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-utilities" (OuterVolumeSpecName: "utilities") pod "23748c76-d92f-46cd-8300-7e80230e3241" (UID: "23748c76-d92f-46cd-8300-7e80230e3241"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.260961 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23748c76-d92f-46cd-8300-7e80230e3241-kube-api-access-4mc4x" (OuterVolumeSpecName: "kube-api-access-4mc4x") pod "23748c76-d92f-46cd-8300-7e80230e3241" (UID: "23748c76-d92f-46cd-8300-7e80230e3241"). InnerVolumeSpecName "kube-api-access-4mc4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.355360 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mc4x\" (UniqueName: \"kubernetes.io/projected/23748c76-d92f-46cd-8300-7e80230e3241-kube-api-access-4mc4x\") on node \"crc\" DevicePath \"\"" Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.355392 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-utilities\") on node \"crc\" DevicePath \"\"" Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.362900 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23748c76-d92f-46cd-8300-7e80230e3241" (UID: "23748c76-d92f-46cd-8300-7e80230e3241"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 13 17:47:31 crc kubenswrapper[4989]: I1213 17:47:31.456252 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23748c76-d92f-46cd-8300-7e80230e3241-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 13 17:47:32 crc kubenswrapper[4989]: I1213 17:47:32.083503 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r9frb" event={"ID":"23748c76-d92f-46cd-8300-7e80230e3241","Type":"ContainerDied","Data":"8754205debe8fccbdb801d6069546d688ad404065775fcd14eb5948c4a6bc3ca"} Dec 13 17:47:32 crc kubenswrapper[4989]: I1213 17:47:32.083569 4989 scope.go:117] "RemoveContainer" containerID="a096ac635950f3a1d015090b3c2c8e921ed26d9c9d68e164bb4a278cf5b2bda7" Dec 13 17:47:32 crc kubenswrapper[4989]: I1213 17:47:32.083607 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r9frb" Dec 13 17:47:32 crc kubenswrapper[4989]: I1213 17:47:32.105964 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r9frb"] Dec 13 17:47:32 crc kubenswrapper[4989]: I1213 17:47:32.107701 4989 scope.go:117] "RemoveContainer" containerID="29a53032f5c82b4787199a306b498f63dc758f87ee690339d8ed5c506db68179" Dec 13 17:47:32 crc kubenswrapper[4989]: I1213 17:47:32.110372 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r9frb"] Dec 13 17:47:32 crc kubenswrapper[4989]: I1213 17:47:32.128014 4989 scope.go:117] "RemoveContainer" containerID="1025a3f12b3cf8e147b74db5867e86eba9d62d1ba7e279349b41c2a5dbaf4a13" Dec 13 17:47:34 crc kubenswrapper[4989]: I1213 17:47:34.024423 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23748c76-d92f-46cd-8300-7e80230e3241" path="/var/lib/kubelet/pods/23748c76-d92f-46cd-8300-7e80230e3241/volumes" Dec 13 17:47:42 crc kubenswrapper[4989]: I1213 17:47:42.014729 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:47:42 crc kubenswrapper[4989]: E1213 17:47:42.015741 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:47:57 crc kubenswrapper[4989]: I1213 17:47:57.015219 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:47:57 crc kubenswrapper[4989]: E1213 17:47:57.016062 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" Dec 13 17:48:10 crc kubenswrapper[4989]: I1213 17:48:10.015248 4989 scope.go:117] "RemoveContainer" containerID="24ac9c0e28e9772c35381e42c1a7a0b4995eb1f5f58922c08bab7abfae85a5c2" Dec 13 17:48:10 crc kubenswrapper[4989]: E1213 17:48:10.016056 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-nh9k2_openshift-machine-config-operator(a2b01148-171a-4f86-84a7-d326739e0dcf)\"" pod="openshift-machine-config-operator/machine-config-daemon-nh9k2" podUID="a2b01148-171a-4f86-84a7-d326739e0dcf" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515117323356024453 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015117323356017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015117317271016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015117317271015462 5ustar corecore